One embodiment is directed generally to a device, and more particularly, to a device that produces haptic effects.
Video games and video game systems have become extremely popular. Video game devices or controllers typically use visual and auditory cues to provide feedback to a user. In some interface devices, kinesthetic feedback (such as active and resistive force feedback) and/or tactile feedback (such as vibration, texture, and heat) is also provided to the user, more generally known collectively as “haptic feedback” or “haptic effects.” Haptic feedback can provide cues that enhance and simplify a user's interaction with a video game controller, or other electronic device. Specifically, vibration effects, or vibrotactile haptic effects, may be useful in providing cues to users of video game controllers or other electronic devices to alert the user to specific events, or provide realistic feedback to create greater sensory immersion within a simulated or virtual environment.
Other devices, such as medical devices, automotive controls, remote controls, and other similar devices where a user interacts with a user input element to cause an action, also benefit from haptic feedback or haptic effects. For example, and not by way of limitation, user input elements on medical devices may be operated by a user outside the body of a patient at a proximal portion of a medical device to cause an action within the patient's body at a distal end of the medical device. Haptic feedback or haptic effects may be employed to alert the user to specific events, or provide realistic feedback to the user regarding an interaction of the medical device with the patient at the distal end of the medical device.
One embodiment is a system that controls a haptic effect experienced at a peripheral device. The system receives a haptic effect definition including haptic data. The system further receives trigger data including at least one of: a position of a trigger of a peripheral device; or a range of the trigger of the peripheral device. The system further determines whether a trigger condition is reached based on the received trigger data. The system further sends a haptic instruction and the haptic effect definition to the peripheral device when the trigger condition is reached. The system further causes a haptic output device (or multiple haptic output devices) to produce haptic effects that are based on the haptic effect definition at the peripheral device in response to the haptic instruction.
Further embodiments, details, advantages, and modifications will become apparent from the following detailed description of the preferred embodiments, which is to be taken in conjunction with the accompanying drawings.
One embodiment is a system that provides haptic feedback that is experienced at a peripheral device, such as a game controller or gamepad. For example, the system can provided a trigger haptic effect that is experienced at a trigger of a controller, or some other peripheral device. The trigger haptic effect can complement rumble haptic effects that are also experienced at the controller, or other peripheral device. As another example, the system can provide a general haptic effect that is experienced at a user input element of the controller, gamepad, or other peripheral device. A wide variety of haptic feedback sensations can be created, such as detents, vibrations, textures, and stiffness. The haptic feedback sensations can complement a game application, or other software application, that can also be executed by the system. The haptic feedback sensations can be suited for a specific genre of game such as: first-person shooter; fantasy/roleplay; racing; or sports. In an alternate embodiment, a peripheral device, such as a game controller or gamepad, can have discrete isolated vibration regions on the handles of the peripheral device.
The system can first receive a haptic effect definition. The system can further receive trigger data, such as a position and/or range of a trigger, or positional information from the trigger, or some other user input element (e.g., joystick, etc.). Examples of such trigger data can include placing the trigger in a specific position, sweeping the trigger through a specific position, placing the trigger within a specific range, or sweeping the trigger through a specific range. The system can modify the haptic effect definition based on the received trigger data. The system can optionally further modify the haptic effect definition based on spatialization data, such as a direction and/or flow of a haptic effect that is based on the haptic effect definition. The system can then cause one or more motors or actuators of the controller, gamepad, or other peripheral device, to play, or otherwise output, haptic feedback based on the modified haptic effect definition, thus causing haptic feedback to be experienced. In other words, the system can cause the controller, gamepad, or other peripheral device, to playback a haptic effect based on the haptic effect definition. As understood by one of ordinary skill in the relevant art, “playback” is the act or instance of reproducing data (e.g., audio data, video data, or haptic data). Thus, in one example, the system can cause an overall trigger haptic effect to be experienced at a trigger of the controller.
In one embodiment, the system can include a number of haptic presets that can be selected, such as trigger haptic effect presets. The system can include a comprehensive library of haptic presets. Each selected haptic effect preset can be modified within a graphical user interface. The modified haptic effect preset can be compiled and executed within an execution of a game application, or other software application, where a controller, gamepad, or other peripheral device, can product a haptic effect, such as a trigger haptic effect.
In another embodiment, the system can include a manual content-driven toolset. The toolset allows a user to design haptic effect definitions alongside audio effect definitions. The system can then encode the haptic effect definitions into haptic files that can have an audio format, or that can be stored alongside audio files. A decoder that resides in a controller, gamepad, or other peripheral device, can subsequently decode the encoded haptic effect definitions. Further, the designed haptic effect definitions can be loaded into a trigger engine and/or a spatialization engine in order to launch haptic effects based on the designed haptic effect definitions. The trigger engine can allow haptic effect launches to be mapped to trigger movement or other trigger behavior, and the spatialization engine can move each haptic effect around to create the sensation of haptic effects originating from a specific direction. The trigger engine and/or the spatialization engine can be made available in an offline tool, where a user can graphically interact with the engines and feel the haptic playback within the controller, gamepad, or other peripheral device, in their hand, once the haptic effects are loaded into the engine.
In yet another embodiment, the system can include an automatic content-driven toolset. The toolset can automatically convert audio effect definitions into haptic effect definitions. The toolset can receive the audio effect definitions as audio files. The toolset can modify one or more parameters to control the conversion of the audio effect definition into a haptic effect definition. As an example, the one or more parameters can include: (a) an audio file parameter that identifies the audio file containing the audio effect definition to be converted; (b) a conversion control parameter that defines how the audio data of the audio effect definition gets divided into two or more distinct frequency bands, which are then applied to two or more distinct outputs (e.g., low rumble motor or actuator; middle rumble motor or actuator; and trigger motor or actuator); (c) a dynamic parameter that controls a magnitude mapping from an audio effect definition to a haptic effect definition, and can adjust a noise floor and dynamics range (either stretching or squashing the dynamics), as desired; and (d) a magnitude (or strength) parameter that controls an ultimate magnitude (or strength) of the haptic effect definition. In an alternate embodiment, rather than receiving an audio effect definition, the toolset can receive a haptic instruction that can be converted into a haptic effect definition. The haptic instruction can take the form of a gameplay parameter received by a gaming application, such as a rate of fire from a weapon within a game that can determine a duration of a haptic effect, a key frame in an animation within a game that can determine a start point and a stop point for the haptic effect, animation data for a directionality on an axis for the haptic effect, a damage amount for a weapon within the game that can determine a magnitude of the haptic effect, etc. Further, the converted haptic effect definition can be loaded into a trigger engine and/or a spatialization engine in order to launch haptic effects based on the designed haptic effect definition, as previously described. The trigger engine and/or the spatialization engine can be made available in an offline tool, as previously described.
In yet another embodiment, the system can include a haptic effect studio engine. A haptic effect studio can be provided to allow a user to define a haptic effect definition for a specific motor or actuator. The designed haptic effect definition can be loaded into a trigger engine and/or a spatialization engine in order to launch haptic effects based on the designed haptic effect definition, as previously described. The trigger engine and/or the spatialization engine can be made available in an offline tool, as previously described.
A computer-readable medium may be any available medium that can be accessed by processor 22 and may include both a volatile and nonvolatile medium, a removable and non-removable medium, a communication medium, and a storage medium. A communication medium may include computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any other form of an information delivery medium known in the art. A storage medium may include RAM, flash memory, ROM, erasable programmable read-only memory (“EPROM”), electrically erasable programmable read-only memory (“EEPROM”), registers, hard disk, a removable disk, a compact disk read-only memory (“CD-ROM”), or any other form of a storage medium known in the art.
In one embodiment, memory 14 stores software modules that provide functionality when executed by processor 22. The modules include an operating system 15 that provides operating system functionality for system 10, as well as the rest of an overall device in one embodiment. The modules further include a haptic trigger control module 16 that controls a haptic effect experienced at a trigger, or some other user input element. In certain embodiments, haptic trigger control module 16 can comprise a plurality of modules, where each module provides specific individual functionality for controlling a haptic effect experienced at a trigger, or some other user input element. System 10 will typically include one or more additional application modules 18 to include additional functionality, such as peripheral firmware which can provide control functionality for a peripheral device, such as a controller 30.
System 10, in embodiments that transmit and/or receive data from remote sources, further includes a communication device 20, such as a network interface card, to provide mobile wireless network communication, such as infrared, radio, Wi-Fi, or cellular network communication. In other embodiments, communication device 20 provides a wired network connection, such as an Ethernet connection or a modem.
System 10 is operably connected to controller 30. Controller 30 is a peripheral device used to provide input to system 10. Controller 30 can be operably connected to system 10 using either a wireless connection or a wired connection. Controller 30 can further include a local processor which can communicate with system 10 using either a wireless connection or a wired connection. Alternatively, controller 30 may be configured to not include a local processor, and all input signals and/or output signals associated with controller 30 can be handled and processed directly by processor 22 of system 10.
Controller 30 can further include one or more digital buttons, one or more analog buttons, one or more bumpers, one or more directional pads, one or more analog or digital sticks, one or more driving wheels, and/or one or more user input elements that can be interacted with by a user, and that can provide input to system 10. Controller 30 can also include one or more analog or digital trigger buttons (or “triggers”) that can further be interacted with by the user, and that can further provide input to system 10. As is described below in greater detail, controller 30 can further include a motor, or another type of actuator or haptic output device, configured to exert a bi-directional push/pull force on at least one trigger of controller 30.
Controller 30 can also include one or more actuators, or other types of haptic output devices. The local processor of controller 30, or, processor 22 in embodiments where controller 30 does not include a local processor, may transmit a haptic signal associated with a haptic effect to at least one actuator of controller 30. The actuator, in turn, outputs haptic effects such as vibrotactile haptic effects, kinesthetic haptic effects, or deformation haptic effects, in response to the haptic signal. The haptic effects can be experienced at a user input element (e.g., a digital button, analog button, bumper, directional pad, analog or digital stick, driving wheel, or trigger) of controller 30. Alternatively, the haptic effects can be experienced at an outer surface of controller 30. The actuator includes an actuator drive circuit. The actuator may be, for example, an electric motor, an electro-magnetic actuator, a voice coil, a shape memory alloy, an electro-active polymer, a solenoid, an eccentric rotating mass motor (“ERM”), a linear resonant actuator (“LRA”), a piezoelectric actuator, a high bandwidth actuator, an electroactive polymer (“EAP”) actuator, an electrostatic friction display, or an ultrasonic vibration generator. An actuator is an example of a haptic output device, where a haptic output device is a device configured to output haptic effects, such as vibrotactile haptic effects, electrostatic friction haptic effects, kinesthetic haptic effects, or deformation haptic effects, in response to a drive signal. In alternate embodiments, the one or more actuators within controller 30 can be replaced by some other type of haptic output device.
Controller 30 can further include one or more speakers. The local processor of controller 30, or, processor 22 in embodiments where controller 30 does not include a local processor, may transmit an audio signal to at least one speaker of controller 30, which in turn outputs audio effects. The speaker may be, for example, a dynamic loudspeaker, an electrodynamic loudspeaker, a piezoelectric loudspeaker, a magnetostrictive loudspeaker, an electrostatic loudspeaker, a ribbon and planar magnetic loudspeaker, a bending wave loudspeaker, a flat panel loudspeaker, a heil air motion transducer, a plasma arc speaker, and a digital loudspeaker.
Controller 30 can further include one or more sensors. A sensor can be configured to detect a form of energy, or other physical property, such as, but not limited to, sound, movement, acceleration, bio signals, distance, flow, force/pressure/strain/bend, humidity, linear position, orientation/inclination, radio frequency, rotary position, rotary velocity, manipulation of a switch, temperature, vibration, or visible light intensity. The sensor can further be configured to convert the detected energy, or other physical property, into an electrical signal, or any signal that represents virtual sensor information, and controller 30 can send the converted signal to the local processor of controller 30, or, processor 22 in embodiments where controller 30 does not include a local processor. The sensor can be any device, such as, but not limited to, an accelerometer, an electrocardiogram, an electroencephalogram, an electromyograph, an electrooculogram, an electropalatograph, a galvanic skin response sensor, a capacitive sensor, a hall effect sensor, an infrared sensor, an ultrasonic sensor, a pressure sensor, a fiber optic sensor, a flexion sensor (or bend sensor), a force-sensitive resistor, a load cell, a LuSense CPS2 155, a miniature pressure transducer, a piezo sensor, a strain gage, a hygrometer, a linear position touch sensor, a linear potentiometer (or slider), a linear variable differential transformer, a compass, an inclinometer, a magnetic tag (or radio frequency identification tag), a rotary encoder, a rotary potentiometer, a gyroscope, an on-off switch, a temperature sensor (such as a thermometer, thermocouple, resistance temperature detector, thermistor, or temperature-transducing integrated circuit), microphone, photometer, altimeter, bio monitor, camera, or a light-dependent resistor.
A housing 102 of controller 100 is shaped to easily accommodate two hands gripping the device, either by a left-handed user or a right-handed user. Those skilled in the art would recognize that controller 100 is merely an example embodiment of a controller of similar shape and size to many “gamepads” currently available for video game console systems, such as a Microsoft® Xbox One™ controller or a PlayStation® DualShock™ controller, and that controllers with other configurations of user input elements, shapes, and sizes may be used, including but not limited to controllers such as a Wii™ remote or Wii™ U Controller, Sony® SixAxis™ controller or Sony® Wand controller, as well as controllers shaped as real life objects (such as tennis rackets, golf clubs, baseball bats, and the like) and other shapes, or controllers with a display or head-mounted display.
Controller 100 includes several user input elements, including an analog or digital stick 110, a button 114, and a trigger 118. As used herein, user input element refers to an interface device such as a trigger, button, analog or digital stick, or the like, which is manipulated by the user to interact with host computer 104. As can be seen in
As can be seen in the block diagram of
Device 500 includes game input management code 501. Game input management code 501 includes a set of computer-readable instructions that manage input provided by controller 520 in the context of a game application, or other type of application, executed within device 500. Device 500 further includes peripheral input application programming interface (“API”) 502. Peripheral input API 502 includes a set of computer-readable functions or routines that allow game input management code 501 to interact with peripheral firmware 510 in order to receive and manage input provided by controller 520. Device 500 further includes rumble API 503. Rumble API includes a set of computer-readable functions or routines that allow game input management code 501 to interact with peripheral firmware 510 in order to transmit rumble instructions to one or more rumble motors, or rumble actuators, of controller 520 (e.g., rumble motors L and R, as illustrated in
Device 500 further includes trigger haptic effect API 504 (identified in
According to the embodiment, trigger haptic effect API 504 can allow game input management code 501 to interact with direct playback/crossover 505, trigger engine 506, and spatialization engine 507, and can further manage direct playback/crossover 505, trigger engine 506, and spatialization engine 507 according to requests invoked by game input management code 501. Further, trigger haptic effect API 504 can store data required for communication with peripheral firmware 510, and required for generation of one or more trigger haptic effects. In an alternate embodiment, trigger haptic effect API 504 can reside within peripheral firmware 510 rather than device 500. Trigger haptic effect API 504 is further described below in greater detail in conjunction with
Device 500 further includes direct playback/crossover 505. Direct playback/crossover 505 receives haptic data as input, produces haptic data as output, and transmits haptic data to one or more targeted motors, or targeted actuators, of controller 520 (e.g., motors L and R, as illustrated in
Device 500 further includes trigger engine 506. Trigger engine 506 can receive haptic data, such as a trigger haptic effect definition, and can modify the haptic data based on data, such as trigger data (e.g., trigger data 513 as illustrated in
Device 500 further includes spatialization engine 507 (identified in
According to the embodiment, spatialization engine 507 can modify the haptic data so that a haptic effect, such as a trigger haptic effect, is scaled for one or more rumble motors, or rumble actuators, of controller 520 (e.g., rumble motors L and R, as illustrated in
Device 500 further includes encoder 508. Encoder 508 encodes haptic data received from direct playback/crossover 505, trigger engine 506, and/or spatialization engine 507 into a format. In one embodiment, the format can be an HES format. Encoder 508 further transmits the encoded haptic data to peripheral firmware 510.
Peripheral firmware 510 includes decoder and crossover 511. Decoder and crossover 511 receives the encoded haptic data from encoder 508 and decodes the encoded haptic data. In certain embodiments, decoder and crossover 511 computes a programmable crossover in order to decode the encoded haptic data. In some of these embodiments, decoder and crossover 511 computes the programmable crossover in real-time. Peripheral firmware 510 further includes trigger control 512. Trigger control 512 is a low-level control API for one or more targeted motors, or targeted actuators, of controller 520 (e.g., motors L and R, as illustrated in
Peripheral firmware 510 further includes trigger data 513. Trigger data 513, as previously described, is data that includes one or more parameters, such as one or more parameters that indicate a position and/or range of one or more triggers of controller 520 (e.g., triggers L and R as illustrated in
Controller 520 includes triggers L and R. Controller 520 further includes gear boxes L and R and motors L and R. Motor L and gearbox L are operably coupled to trigger L within controller 520. Likewise, motor R and gearbox R are operably coupled to trigger R within controller 520. When motor L receives a trigger instruction, motor L and gearbox L collectively cause a trigger haptic effect to be experienced at trigger L. Likewise, when motor R receives a trigger instruction, motor R and gearbox R collectively cause a trigger haptic effect to be experienced at trigger R. According to the embodiment, peripheral firmware 510 sends trigger instructions to motors L and R of controller 520 using drive electronics 530. Controller 520 further includes potentiometers L and R. Potentiometer L can detect a position and/or range of trigger L, and can further send the detected position and/or range of trigger L to peripheral firmware 510 as trigger data. Likewise, potentiometer R can detect a position and/or range of trigger R, and can further send the detected position and/or range of trigger R to peripheral firmware 510 as trigger data. In one embodiment, potentiometers L and R can each be replaced with another type of position sensor, such as a hall effect sensor. Controller 520 further includes rumble motors L and R. When rumble motor L receives a rumble instruction, rumble motor L causes a haptic effect to be experienced along a left body of controller 520. Likewise, when rumble motor R receives a rumble instruction, rumble motor R cause a haptic effect to be experienced along a right body of controller 520. According to the embodiment, peripheral firmware 510 sends rumble instructions to rumble motors L and R of controller 520 using rumble drive electronics 530.
In an alternate embodiment, one or more targeted motors, or targeted actuators, can be operably coupled to one or more user input elements (such as one or more digital buttons, one or more analog buttons, one or more bumpers, one or more directional pads, one or more analog or digital sticks, one or more driving wheels) of controller 520. According to the alternate embodiment, peripheral firmware 510 can sends instructions to the one or more targeted motors or targeted actuators, causing the one or more targeted motors or targeted actuators to produce haptic effects that are experienced at the one or more user input elements of controller 520.
User interface 600 further includes effect definitions 630. According to the embodiment, the user can save a modified haptic effect definition as a new haptic effect definition, where the new haptic effect definition is displayed within effect definitions 630. The new haptic effect definition can be stored within a haptic file or haptic stream. In one embodiment, a format of the haptic file or haptic stream can be an HES format. The new haptic effect definition can further be exported to an external haptic file or external haptic stream. User interface 600 further includes a play button 640. Interacting with play button 640 can cause the system to output a haptic effect at a controller that can be operably controlled to user interface 600. If the haptic effect is a trigger haptic effect, the system can output the trigger haptic effect at a trigger of the controller. The haptic effect can be a selected pre-defined haptic effect definition or a selected new haptic effect definition.
User interface 600 further includes trigger engine area 650. Trigger engine area 650 is an editable visual area that can visualize a trigger haptic effect that is generated by a trigger engine (such as trigger engine 506 of
The crossover input warp algorithm can reside either in the device itself, or reside on the opposite side of a communications link, executing on a processor different from that of the device. The crossover input warp algorithm may also separate the input data (haptic or audio) into two bands, where lower frequencies are separated and then optionally further transformed before being applied to one or more actuator outputs, and higher frequencies are separated and then optionally transformed before being applied to a number of actuators distinct from those used for the lower-frequency separated data. This type of data separation could occur on an arbitrary number of frequency bands and actuator outputs. In alternate embodiments, the input data (audio or haptic) can be separated into multiple overlapping frequency regions, which are then each optionally transformed and applied to a number of output actuators. Another set of embodiments could create a number of signal strength bands, where the input data (audio or haptic) is separated according to output power or strength (such as through peak detection, RMS calculations, etc.), and these separated data streams are each applied to one or more distinct sets of actuators. In alternate embodiments, the input data (audio or haptic) can be separated according to output power or strength (such as through peak detection, RMS calculations, etc.) into distinct but overlapping data streams, instead of completely distinct streams, where the strength filtering algorithms capture overlapping regions of strength, optionally apply the transformations and apply each of the outputs to a number of output actuators.
The system can further send the encoded audio effect definition or the encoded haptic effect definition to a human interface device (“HID”) interpreter 740 that resides on controller 750. HID interpreter 740 receives and interprets the encoded audio effect definition or the encoded haptic effect definition in order to provide a trigger haptic effect at a trigger of controller 750. In one embodiment, a system can further modify the encoded audio effect definition or the encoded haptic effect definition using a trigger engine (such as trigger engine 506 of
Once a user of the system has authored a trigger haptic effect using audio authoring component 800, the user can preview the trigger haptic effect. The preview functionality can allow for further customization of the trigger haptic effect. Upon previewing the trigger haptic effect, the system can send the authored audio effect definition to four-channel output driver 801, where four-channel output driver 801 can stream the audio effect definition as four channels of audio data. In one embodiment, four-channel output driver 801 can be a four-channel ASIO output driver. In an alternate embodiment, four-channel output driver 801 can be replaced by another driver that streams the audio effect definition as any plural number of channels of audio data, such as six or eight channels of audio data.
Further, the system can send the audio stream to audio-to-haptic converter 802, where audio-to-haptic converter 802 can convert the audio effect definition of the audio stream into a haptic effect definition using a haptic conversion algorithm. In one embodiment, each separate channel of the audio effect definition that corresponds to a motor, or actuator, can be converted into a channel of a haptic effect definition. Example haptic conversion algorithms are described in the following patents or patent applications (all of which are hereby incorporated by reference in their entirety): U.S. Pat. Nos. 7,979,146; 8,000,825; 8,378,964; U.S. Pat. App. Pub. No. 2011/0202155; U.S. Pat. App. Pub. No. 2011/0215913; U.S. Pat. App. Pub. No. 2012/0206246; U.S. Pat. App. Pub. No. 2012/0206247; U.S. Pat. App. Pub. No. 2013/0265286; U.S. Pat. App. Pub. No. 2013/0131851; U.S. Pat. App. Pub. No. 2013/0207917; U.S. Pat. App. Pub. No. 2013/0335209; U.S. Pat. App. Pub. No. 2014/0064516; U.S. patent application Ser. Nos. 13/661,140; 13/785,166; 13/788,487; 14/078,438; 14/078,442; 14/078,445; 14/051,933; 14/020,461; 14/020,502; 14/277,870; and 14/467,184.
The system can further send the converted haptic effect definition to HES multi-channel encoder 803, where multi-channel encoder 803 can encode the converted haptic effect definition into an external format, such as an HES format. The system can further send the encoded and converted haptic effect definition to trigger controller interface (“I/F”) 804 that resides on controller 805. Trigger controller I/F 804 can receive and interpret the encoded and converted haptic effect definition in order to preview the authored trigger haptic effect at a trigger of controller 805.
In this embodiment, the system can provide audio authoring component 810, where audio authoring component 810 is identical to audio authoring component 800. Once a user of the system has authored a trigger haptic effect using audio authoring component 810, the user can save the trigger haptic effect. Upon saving the trigger haptic effect, the system can export the audio effect definition as separate audio files 811. In one embodiment where the audio effect definition includes four channels, audio files 811 can include four audio files. In an alternate embodiment, where the audio effect definition includes another number of channels, audio files 811 can include that number of separate audio files. In certain embodiments, audio files 811 can be a Waveform Audio File (“WAV”) format. The system can further send audio files 811 to a HES encoder graphical user interface (“GUI”) 812, where HES encoder GUI 812 can encode audio files 811 into a single audio file. In one embodiment, the audio file can be an HES format. Further, the system can send the audio file to audio-to-haptic converter 812, where audio-to-haptic converter 813 can convert the audio effect definition of the audio file into a haptic effect definition using a haptic conversion algorithm. In one embodiment, each separate channel of the audio effect definition that corresponds to a motor, or actuator, can be converted into a channel of a haptic effect definition. The system can further send the converted haptic effect definition to HES multi-channel encoder 814, where multi-channel encoder 814 can encode the converted haptic effect definition into an external format, such as an HES format. The system can further store the encoded and converted haptic effect definition within a haptic file 815. In one embodiment, haptic file 815 can be an HES file.
Once a user of the system has authored a trigger haptic effect using audio authoring component 900, the user can preview the trigger haptic effect. Upon previewing the trigger haptic effect, the system can send the authored audio effect definition to single-channel output driver 901, where single-channel output driver 901 can stream the audio effect definition as a single channel of audio data. In one embodiment, single-channel output driver 901 can be a single-channel ASIO output driver. Further, the system can send the audio stream to audio-to-haptic converter 902, where audio-to-haptic converter 902 can convert the audio effect definition of the audio stream into a haptic effect definition using a haptic conversion algorithm. In one embodiment, each separate channel of the audio effect definition that corresponds to a motor, or actuator, can be converted into a channel of a haptic effect definition. Even further, the system can send the converted haptic effect definition to crossover GUI 905, where crossover GUI 905 can apply a crossover input warp algorithm to separate the converted haptic effect definition into three different channels that can be mapped to three different outputs (such as: (1) a low-frequency rumble motor, or rumble actuator; (2) a medium-frequency rumble motor, or rumble actuator; or (3) a high-frequency targeted motor, or targeted actuator).
The system can further send the converted haptic effect definition to HES multi-channel encoder 903, where multi-channel encoder 903 can encode the converted haptic effect definition into an external format, such as an HES format. The system can further send the encoded and converted haptic effect definition to trigger controller I/F 904 that resides on controller 906. Trigger controller I/F 904 can receive and interpret the encoded and converted haptic effect definition in order to preview the authored trigger haptic effect at a trigger of controller 906.
In this embodiment, the system can provide audio authoring component 910, where audio authoring component 910 is identical to audio authoring component 900. Once a user of the system has authored a trigger haptic effect using audio authoring component 910, the user can save the trigger haptic effect. Upon saving the trigger haptic effect, the system can export the audio effect definition as a single audio file 911. In certain embodiments, audio file 911 can be a WAV format. The system can further export crossover settings 912. The system can further send audio file 911 to a HES encoder GUI 913, where HES encoder GUI 913 can encode audio file 911 and crossover settings 912 into a single audio file. In one embodiment, the audio file can be an HES format. The system can further send the audio file to HES single-channel and crossover encoder 914, where single-channel and crossover encoder can encode the audio file into an external format, such as an HES format. The system can further store the encoded audio file within a haptic file 915. In one embodiment, haptic file 915 can be an HES file.
According to the embodiment, the system can send the four channels of the haptic effect definition included within haptic file 1000 to strength control 1010, where strength control 1010 can modify a strength, or magnitude, of the haptic data included within each channel of the haptic effect definition. The system can then send the four channels of the haptic effect definition to front/back (“F/B”) spatialization 1020, where F/B spatialization 1020 can modify the haptic data included within each channel of the haptic effect definition based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a frontwards or backwards direction. Further, spatialization data can include one or more hand positions. According to the embodiment, F/B spatialization 1020 can modify the haptic data included within each channel so that a haptic effect is scaled for each motor, or actuator. The system can then send channel LR to low rumble motor 1030 (identified in
The system can further send channels LT and RT to left/right (“L/R”) spatialization 1050, where L/R spatialization 1050 can modify the haptic data included within channels LT and RT based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a left or right direction. Further, spatialization data can include one or more hand positions. According to the embodiment, L/R spatialization 1050 can modify the haptic data included within each channel so that a haptic effect is scaled for each motor, or actuator. The system can then send channel LT to left trigger targeted motor 1060 (identified in
According to the embodiment, the system can send the channel of the haptic effect definition included within haptic file 1100, and the one or more crossover parameters also included within haptic file 1100, to programmable crossover 1110. Programmable crossover 1110 can apply a crossover input warp algorithm using the one or more crossover parameters to separate the channel into three different channels: a low-frequency channel; a medium-frequency channel; and a high-frequency channel. The low-frequency channel includes a portion of the haptic data included within the haptic effect definition that includes one or more low frequencies. The medium-frequency channel includes a portion of the haptic data included within the haptic effect definition that includes one or more medium frequencies. The high-frequency channel includes a portion of the haptic data included within the haptic effect definition that includes one or more high frequencies.
The system can then send the three channels of the haptic effect definition to F/B spatialization 1120, where F/B spatialization 1120 can modify the haptic data included within each channel of the haptic effect definition based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a frontwards or backwards direction. Further, spatialization data can include one or more hand positions. According to the embodiment, F/B spatialization 1120 can modify the haptic data included within each channel so that a haptic effect is scaled for each motor, or actuator. The system can then send the low frequency channel to low rumble motor 1130 (identified in
The system can further send the high-frequency channel to L/R spatialization 1150, where L/R spatialization 1150 can modify the haptic data included within the high frequency channel based on spatialization data. In one embodiment, the direction and/or flow of the haptic effect can be a left or right direction. Further, spatialization data can include one or more hand positions. According to the embodiment, L/R spatialization 1150 can modify the haptic data included within the channel so that a haptic effect is scaled for each motor, or actuator. The system can then send the high-frequency channel to left trigger targeted motor 1160 (identified in
According to the embodiment, the system can send the four channels of the audio effect definition included within audio file 1200 to audio-to-haptic converter 1210, where audio-to-haptic converter 1210 can convert the audio effect definition into a haptic effect definition using a haptic conversion algorithm. In one embodiment, each separate channel of the audio effect definition can be converted into a channel of a haptic effect definition. In the illustrated embodiment: channel LR can be converted using a peak/decimation filter with a range of less than 60 hertz (“Hz”); channel MR can be converted using a peak/decimation filter with a value of 60 Hz; and channels LT and RT can each be converted using a peak/decimation filter with a range of 200 Hz-2 kHz.
The system can further send the four channels of the converted haptic effect definition to encoder/decoder 1220, where encoder/decoder 1220 can encode each channel of the converted haptic effect definition into an external format, such as an HES format. The system can then send the four encoded channels of the converted haptic effect definition to F/B spatialization 1230, where F/B spatialization 1230 can modify the converted haptic data included within each encoded channel of the converted haptic effect definition based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a frontwards or backwards direction. Further, spatialization data can include one or more hand positions. According to the embodiment, F/B spatialization 1230 can modify the converted haptic data included within each encoded channel so that a haptic effect is scaled for each motor, or actuator. The system can then send encoded channel LR to low rumble motor 1240 (identified in
The system can further send encoded channels LT and RT to L/R spatialization 1260, where L/R spatialization 1260 can modify the converted haptic data included within encoded channels LT and RT based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a left or right direction. Further, spatialization data can include one or more hand positions. According to the embodiment, L/R spatialization 1260 can modify the haptic data included within each channel so that a haptic effect is scaled for each motor, or actuator. The system can then send channel LT to left trigger targeted motor 1270 (identified in
According to the embodiment, the system can send the channel of the audio effect definition included within audio file 1300, and, in one embodiment, the one or more crossover parameters also included within audio file 1300, to programmable crossover 1310. Programmable crossover 1310 can apply a crossover input warp algorithm (in one embodiment, using the one or more crossover parameters) to separate the channel into three different channels: a low-frequency channel; a medium-frequency channel; and a high-frequency channel. Programmable crossover 1310 can further convert the audio effect definition into a haptic effect definition using a haptic conversion algorithm. In one embodiment, each separate channel of the audio effect definition can be converted into a channel of a haptic effect definition. In the illustrated embodiment: the low-frequency channel can be converted using a peak/decimation filter with a range of less than 60 hertz (“Hz”); the medium-frequency channel can be converted using a peak/decimation filter with a value of 60 Hz; and the high-frequency channel can each be converted using a peak/decimation filter with a range of 200 Hz-2 kHz.
The system can further send the three channels of the converted haptic effect definition to encoder/decoder 1320, where encoder/decoder 1320 can encode each channel of the converted haptic effect definition into an external format, such as an HES format. The system can then send the three channels of the haptic effect definition to F/B spatialization 1330, where F/B spatialization 1330 can modify the haptic data included within each channel of the haptic effect definition based on spatialization data. The spatialization data can include a direction and/or flow of a haptic effect. In one embodiment, the direction and/or flow of the haptic effect can be a frontwards or backwards direction. Further, spatialization data can include one or more hand positions. According to the embodiment, F/B spatialization 1330 can modify the haptic data included within each channel so that a haptic effect is scaled for each motor, or actuator. The system can then send the low-frequency channel to low rumble motor 1340 (identified in
The system can further send the high-frequency channel to L/R spatialization 1360, where L/R spatialization 1360 can modify the haptic data included within the high-frequency channel based on spatialization data. In one embodiment, the direction and/or flow of the haptic effect can be a left or right direction. Further, spatialization data can include one or more hand positions. According to the embodiment, L/R spatialization 1360 can modify the haptic data included within the channel so that a haptic effect is scaled for each motor, or actuator. The system can then send the high frequency channel to left trigger targeted motor 1370 (identified in
Using user interface 1400, a user can create one or more trigger definitions, where a trigger definition defines a condition that causes a trigger haptic effect to be generated or modified. In some embodiments, a condition can be a position of a trigger, where the position is within trigger input range 1420. In other embodiments, a condition can be a range of the trigger, where the range is within trigger input range 1420. User interface 1400 includes threshold trigger definition 1430, which is an example of a trigger definition. Threshold trigger definition 1430 causes a trigger haptic effect that is based on a specified trigger haptic effect definition to be generated when the trigger reaches a specified position, where the specified position is 99 in the illustrated embodiment. Threshold trigger definition 1430 can define that the trigger haptic effect be generated only on a press of the trigger, only on a release of the trigger, or both a press and a release of the trigger. In an alternate embodiment, threshold trigger definition 1430 can be replaced with a position trigger definition that causes a trigger haptic effect that is based on a specified trigger haptic effect definition to only be generated when the trigger resides at a specified position, as opposed to simply reaching the specified position. User interface 1400 further includes range trigger definition 1440, which is another example of a trigger definition. Range trigger definition 1440 causes a trigger haptic effect that is based on a specified trigger haptic effect definition to be generated when the trigger reaches a specified range, where the specified range is 164 to 255 in the illustrated embodiment.
User interface 1500 includes flow 1510. Flow 1510 allows a user to programmatically manage a flow of a trigger haptic effect. A flow is a temporal start-of-playback offset modification to delay playback on individual targeted motors, targeted actuators, rumble motors, or rumble actuators of a controller. Alternatively, a flow can be a duration modification to modify a duration of a haptic effect experienced at targeted motors, targeted actuators, rumble motors, or rumble actuators of a controller. For example, a flow can be defined so that haptic playback first begins on a left targeted motor or targeted actuator, then subsequently begins on a middle rumble motor or rumble actuator, and then further begins on a right targeted motor or targeted actuator. In this example, a flow of the overall trigger haptic effect is left-to-right, as a user of a controller first experiences the haptic playback of the overall trigger haptic effect at the left of the controller, then at the middle of the controller, and then at the right of the controller. A flow can be from left to right or vice-versa, front to back or vice-versa, or a combination of the two. Thus, a flow can define a haptic playback vector. Flow 1510 can be visualized within user interface 1500 as an arrow that can be placed horizontally, vertically, or diagonally within user interface 1500. Thus, by interacting with flow 1510, a user can modify one or more delays applied to various motors or actuators of the controller to stagger haptic playback.
User interface 1500 further includes direction 1520. Direction 1520 allows a user to programmatically modify a direction of a trigger haptic effect. A direction is a magnitude (or strength) modification to emphasize a front-back and/or left-right bias (or balance) among various motors or actuators of a controller. Alternatively, a direction can be a frequency modification. For example, a direction can be defined so that haptic playback of the trigger haptic effect is the strongest at the right of the controller. Direction 1520 can be visualized within user interface 1500 as a point within a two-dimensional grid or space defined by two axes. Thus, by interacting with direction 1520, a user can modify magnitudes (or strengths) applied to various motors or actuators to emphasize a left-right and/or front-back bias (or balance).
User interface 1500 further includes strength 1530. Strength 1530 allows a user to programmatically modify a magnitude (or strength) of an overall trigger haptic effect either before or during playback. Strength 1530 can be visualized within user interface 1500 as a slider. Thus, by interacting with strength 1530, a user can modify an overall magnitude (or strength) of a trigger haptic effect. User interface 1500 further includes play speed 1540. Play speed 1540 allows a user to programmatically modify a play speed, or rate, at which a system (such as system 10 of
According to the embodiment, trigger haptic effect API 1600 can be accessed by application 1610, which is a software application, such as a game application, that can be executed on a system (such as system 10 of
In one embodiment, application 1610 can access device definition 1601 to acquire a target hardware device (i.e., HW 1640) where a haptic effect is to be played. By accessing device definition 1601, application 1610 can further access timer definition 1602, trigger definition 1603, and protocol definition 1604. Application 1610 can further access haptic effect definition 1621 from effect library 1620 to instantiate a haptic effect. Application 1610 can further cause the haptic effect be played at the target hardware device (i.e., HW 1640) by sending an instruction to the target hardware device (i.e., HW 1640) via trigger haptic effect API 1600 and FW 1630.
The architecture further includes trigger engine 1730. As previously described, trigger engine 1730 can receive a trigger haptic effect definition and can modify the trigger haptic effect definition based on trigger data, such as a position and/or range of a trigger of a controller. The architecture further includes trigger hardware interface 1740 (identified in
One type of a built-in haptic effect definition is static haptic effect definition 2011 (identified in
Another type of a built-in haptic effect definition is dynamic haptic effect definition 2012 (identified in
Another type of a built-in haptic effect definition is direct control haptic effect definition 2013 (identified in
The system further includes core effect library 2020 (identified in
The system further includes low-level API 2030. Low-level API 2030 can receive an instruction to play a haptic effect based on a haptic effect definition, and can convert the instruction to a low-level instruction that can be interpreted by a controller 2040. An example of low-level API 2030 is Xbox® API 2031 by Microsoft Corporation, and an example of controller 2040 is Xbox® controller 2041 by Microsoft Corporation.
User interface 2100 further includes timeline 2130. According to the embodiment, a user can select a haptic effect preset displayed within open effects 2110, and timeline 2130 can display a graphical representation of the haptic effect definition that is represented by the selected haptic effect preset. In the illustrated embodiment, the haptic effect definition includes four channels, with each channel including haptic data that is mapped for a specific output (e.g., (1) targeted motor or actuator for a right trigger; (2) targeted motor or actuator for a left trigger; (3) right rumble motor or actuator; and (4) left rumble motor or actuator), and each channel being displayed along the timeline. However, in other embodiments, the haptic effect definition can include any number of channels. Further, a user can modify one or more channels of the selected haptic effect definition by interacting with one or more display elements within timeline 2130. By modifying one or more channels of a haptic effect definition, one can modify one or more attributes of a corresponding haptic effect.
User interface 2100 further includes effect properties 2140. Effect properties 2140 is an editable visual area that can visualize a trigger haptic effect that is generated by a trigger engine (such as trigger engine 506 of
User interface 2100 further includes spatialization 2150. Spatialization 2150 is an editable visual area that can visualize a trigger haptic effect that is originally generated by the trigger engine and further modified by a spatialization engine (such as spatialization engine 507 of
User interface 2400 includes plotter 2401. Plotter 2401 takes a haptic effect definition specified by a user as input, and sends the haptic data includes within the haptic effect definition through adapter layer 2410 to trigger API layer 2420. Trigger API layer 2420 sends back individual channel data that plotter 2401 displays within user interface 2400. Render 2402 takes input from controller GUI 2403 and starts a haptic player render loop. The input is routed through adapter layer 2410, which has callbacks setup with trigger API layer 2420 to and relay controller input 2413 (such as button and trigger input) sent from controller 2414. Adapter layer 2410 can also communicate with plotter 2401 while the render loop is running to update user interface 2400. Controller GUI 2403 can also select controller 2414 using controller selector 2412, and can show what is connected. Controller GUI 2403 can also set up a trigger activation point. Further, importer/exporter 2404 can take input audio files and convert them to a haptic file. In one embodiment, an audio file is a WAV file. Further, adapter layer 2410 can be embedded within user interface 2400, or can be a separate library. When adapter layer 2410 is a separate library, adapter layer 2410 can be a separate C++ library.
The system further includes haptic engine 2510. Haptic engine 2510 is a high-level API that can utilize a low level API to perform the playing of a haptic effect, and to add haptic effects to game application 2500. Haptic engine 2510 can load, start, stop, and render a haptic effect. Haptic engine 2510 can interface with haptic effect parser 2520 to parse/get information about a haptic effect. Haptic engine 2510 can further interface with haptic mixer 2530 to start or stop an effect and modify a mixer buffer. Haptic engine 2510 can further interface with haptic device handler 2550 to get a device handle of, and render haptic effects on, a controller, gamepad, or other peripheral device.
The system further includes haptic effect parser 2520. Haptic effect parser 2520 includes an API that can load a haptic effect in memory, verify its format, and obtain information about the haptic effect, such as size, duration, and haptic data. The system further includes haptic mixer 2530. Haptic mixer 2530 supports playback of multiple haptic effects at the same time. The system further includes haptic device handler 2540. Haptic device handler 2540 can initiate and manage communication with a controller, gamepad, or other peripheral device. Haptic device handler 2540 can interface with a Universal Serial Bus (“USB”) communication layer and obtain a device handle of the controller, gamepad, or other peripheral device. Haptic device handler 2540 can further initialize several state machine structures critical for haptic effect playback.
The system further includes trigger haptic report handler 2550. Trigger haptic report handler 2550 can package haptic data into USB HID packets according to a trigger communication protocol. The system further includes platform compliant USB HID library 2560. Platform compliant USB HID library 2560 includes one or more computer-readable routines to interface with USB HID and Bluetooth HID class of controllers, gamepads, or other peripheral devices. The system further includes peripheral firmware 2570 (identified in
In one embodiment, a controller, gamepad, or other peripheral device, can have a customized protocol for conveying haptic data and for driving individual motors or actuators. Accordingly, an audio driver can be provided that receives an audio file that includes a haptic effect authored as an audio effect definition from an audio authoring component, and that sends the audio data included within the audio file to the controller, gamepad, or other peripheral device. In one embodiment, the audio authoring component can be a “Pro Tools®” product by Avid Technology, Inc. The audio driver can get loaded during a boot up process. The audio driver can expose a necessary number of audio channels in order to make haptic effect definitions possible for using all the motors or actuators in the controller, gamepad, or other peripheral device. The audio driver can further work in user space, and can be accessible to all user space audio editing/playback applications. The audio driver can further read the audio data that an audio authoring component sends to the controller, gamepad, or other peripheral device. The audio driver can further perform necessary processing on the audio data being presented and can convert the audio data into haptic data, such as actuator drive values. The audio driver can further communicate the haptic data to the controller, gamepad, or other peripheral device over a communication interface.
According to the embodiment, a controller, gamepad, or other peripheral device, can include four actuators. Two actuators can be used as trigger actuators influencing haptic feedback on triggers. The trigger actuators can be bi-directional. Two kinds of direction events can happen with the trigger actuators: PUSH and PULL. The PUSH and PULL directions can be relative to a user's finger on the trigger. Two other actuators can be used as rumble actuators influencing general haptic feedback or rumble feedback within the controller, gamepad, or other peripheral device. The rumble actuators can be uni-directional. More specifically, the rumble actuators can spin in either a clockwise direction or a counter-clockwise direction, but not both directions. The direction of the motion can be dependent on the controller and/or the drive electronics of the controller.
In this embodiment, the following channel layout can be chosen for the audio driver:
In one embodiment, an audio format chosen for a 16-bit PCM can be 44.1 KHz. The audio driver can receive the audio data from an audio authoring component, convert the audio data into haptic data (e.g., drive values), and communicate the haptic data to the controller accordingly.
(PushChannelDecimatedValue−PullChannelDecimatedValue)→Scale it to [0,255] Drive value for triggers:
(DecimatedValue)→Scale it to [128,255] Drive value for rumbles:
Subsequently, trigger protocol packet manager 2835 obtains drive values for all the actuators (e.g., all four actuators) and packages the drive values as data packets, such as USB HID packets, according to a trigger communication protocol. Further, XPC handler 2845 receives the data packets from trigger protocol packet manager 2835 and sends the data packets to XPC service 2810, which is a background service. At 2855, XPC service 2810 receives the data packets and, at 2865, sends the data packets to 2865 to a controller 2820 (identified in
The flow begins and proceeds to 3110. At 3110, an audio effect definition including audio data is converted into a haptic effect definition including haptic data. In one embodiment, the haptic effect definition can define a haptic effect that can be experienced at a user input element of a peripheral device, or that can be experienced at the peripheral device. Further, in one embodiment, a user input element can be a trigger, and a peripheral device can be a controller or gamepad. The flow then proceeds to 3120. At 3120, the haptic effect definition is received. In one embodiment, the haptic effect definition can be a trigger haptic effect definition. Further, in an embodiment, the haptic effect definition can recreate a haptic effect of a different format. The flow then proceeds to 3130.
At 3130, the haptic data of the haptic effect definition is modified based on received spatialization data. The spatialization data can include at least one of: a direction of the haptic effect; or a flow of the haptic effect. In certain embodiments, the modifying the haptic data of the haptic effect definition can include scaling at least one of: a magnitude of the haptic data; a frequency of the haptic data; or a duration of the haptic data. The flow then proceeds to 3140. At 3140, trigger data is received. The trigger data can include at least one of: a position of the trigger of the peripheral device; or a range of the trigger of the peripheral device. The flow then proceeds to 3150.
At 3150, it is determined whether a trigger condition is reached based on the received trigger data. The flow then proceeds to 3160. At 3160, the haptic data of the haptic effect definition is encoded. In certain embodiments, the haptic data of the haptic effect definition can be encoded within an audio file or audio stream. The flow then proceeds to 3170.
At 3170, a haptic instruction and the haptic effect definition are sent to the peripheral device. In certain embodiments, the haptic instruction and the haptic effect definition can be sent to the peripheral device when the trigger condition is reached. In these embodiments, the haptic instruction can be a trigger instruction. In certain embodiments, the sending the haptic effect definition to the peripheral device can include directly sending the haptic data of the haptic effect definition to the peripheral device. In other embodiments, the sending the haptic effect definition to the peripheral device can include: (1) converting the haptic data of the haptic effect definition from a first format to a second format; and (2) sending the converted haptic data of the haptic effect definition to the peripheral device. The flow then proceeds to 3180.
At 3180, the encoded haptic data of the haptic effect definition is decoded. In certain embodiments, the encoded haptic data can be decoded by computing a programmable crossover. The flow then proceeds to 3190. At 3190, the haptic instruction causes a haptic output device to produce a haptic effect based on the haptic effect definition at the peripheral device. In certain embodiments, the haptic instruction can cause the haptic output device to produce the haptic effect based on the haptic effect definition at the user input element of the peripheral device. In certain embodiments, the haptic instruction can be a trigger instruction, the haptic effect definition can be a trigger haptic effect definition, the haptic output device can be a targeted haptic output device, and the targeted haptic output device can produce the haptic effect at a trigger of the peripheral device. In certain embodiments, the targeted haptic output device can be a targeted actuator. In some of those embodiments, the targeted actuator can be a targeted motor. Further, in certain embodiments, the haptic instruction can cause multiple haptic output devices to produce multiple haptic effects based on the haptic effect definition at the peripheral device. The flow then ends.
In an embodiment, as previously described, a spatialization engine can receive haptic data, such as a trigger haptic effect definition, and can modify the haptic data based on spatialization data, where spatialization data can include one or more parameters. Thus, the spatialization engine can localize or spatialize haptic effects. More specifically, the spatialization engine can produce a haptic effect that conveys a distance of the haptic effect by scaling or attenuating the haptic effect on an actuator or motor based on the distance of the haptic effect. The spatialization engine can further produce a haptic effect that conveys movement on a controller, gamepad, or other peripheral device by delaying or scaling the haptic effect on different actuators or motors. The spatialization engine can be a component of an API or library, or can be implemented in firmware for a controller, gamepad, or other peripheral device.
Device 3200 includes effect library 3201, where effect library 3201 can include one or more haptic effect definitions. In the embodiment, these haptic effect definitions can be identified as unspatialized haptic effect definitions, as they are haptic effect definitions that have not been modified by a spatialization engine. Device 3200 further includes game 3202, where game 3202 is a software application, such as a game application, that can be executed on the system. According to the embodiment, game 3202 can generate one or more spatialization parameters, where the one or more spatialization parameters can define a position, velocity, direction, and/or flow of a haptic effect defined by a haptic effect definition that is stored within effect library 3201.
Device 3200 further includes spatialization engine 3203 (identified in
Device 3300 includes effect library 3301, where effect library 3301 can include one or more haptic effect definitions, identified as unspatialized haptic effect definitions. Device 3300 further includes game 3302, where game 3302 is a software application, such as a game application, that can be executed on the system. According to the embodiment, game 3302 can generate one or more spatialization parameters, where the one or more spatialization parameters can define a position, velocity, direction, and/or flow of a haptic effect defined by a haptic effect definition that is stored within effect library 3301.
Controller 3310 includes spatialization engine 3311 (identified in
Thus, in one embodiment, a system can provide a haptic control architecture that can generate a haptic effect that is experienced at a peripheral device, such as a controller or gamepad. The haptic effect can be a trigger haptic effect that is experienced at a trigger of the peripheral device. The trigger haptic effect can be customized by the haptic control architecture based on trigger data that is received by the system, where the trigger data can include a position and/or range of the trigger. The trigger haptic effect can be further spatialized by the haptic control architecture so that the trigger haptic effect is scaled at each motor or actuator of the peripheral device, so that the trigger haptic effect includes a sense of directionality and/or flow. By incorporating haptic feedback experienced at a peripheral device, and in particular, haptic feedback experienced at a trigger of the peripheral device, into a gaming application that is executed by the system, a more realistic and immersive gaming experience can be provided.
The features, structures, or characteristics of the invention described throughout this specification may be combined in any suitable manner in one or more embodiments. For example, the usage of “one embodiment,” “some embodiments,” “certain embodiment,” “certain embodiments,” or other similar language, throughout this specification refers to the fact that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the present invention. Thus, appearances of the phrases “one embodiment,” “some embodiments,” “a certain embodiment,” “certain embodiments,” or other similar language, throughout this specification do not necessarily all refer to the same group of embodiments, and the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
One having ordinary skill in the art will readily understand that the invention as discussed above may be practiced with steps in a different order, and/or with elements in configurations which are different than those which are disclosed. Therefore, although the invention has been described based upon these preferred embodiments, it would be apparent to those of skill in the art that certain modifications, variations, and alternative constructions would be apparent, while remaining within the spirit and scope of the invention. In order to determine the metes and bounds of the invention, therefore, reference should be made to the appended claims.
This application is a continuation of U.S. patent application Ser. No. 14/539,111, filed on Nov. 12, 2014, the specification of which is hereby incorporated by reference. U.S. patent application Ser. No. 14/539,111 claims priority of U.S. Provisional Patent Application Ser. No. 61/904,342, filed on Nov. 14, 2013, the disclosure of which is hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
5473126 | Wu | Dec 1995 | A |
5959613 | Rosenberg et al. | Sep 1999 | A |
6005551 | Osborne et al. | Dec 1999 | A |
6169540 | Rosenberg et al. | Jan 2001 | B1 |
6184868 | Shahoian | Feb 2001 | B1 |
6343349 | Braun et al. | Jan 2002 | B1 |
6424356 | Chang et al. | Jul 2002 | B2 |
6468158 | Ootori et al. | Oct 2002 | B1 |
6563487 | Martin et al. | May 2003 | B2 |
6710764 | Burgel et al. | Mar 2004 | B1 |
7182691 | Schena | Feb 2007 | B1 |
7194287 | Usuki et al. | Mar 2007 | B2 |
7196688 | Schena | Mar 2007 | B2 |
7336260 | Martin et al. | Feb 2008 | B2 |
7791588 | Tierling et al. | Sep 2010 | B2 |
7808488 | Martin et al. | Oct 2010 | B2 |
7965276 | Martin | Jun 2011 | B1 |
7979146 | Ullrich et al. | Jul 2011 | B2 |
7979797 | Schena | Jul 2011 | B2 |
8000825 | Ullrich et al. | Aug 2011 | B2 |
8159461 | Martin et al. | Apr 2012 | B2 |
8164573 | Da Costa et al. | Apr 2012 | B2 |
8210942 | Shimabukuro et al. | Jul 2012 | B2 |
8378964 | Ullrich et al. | Feb 2013 | B2 |
8384316 | Houston et al. | Feb 2013 | B2 |
8723820 | Han | May 2014 | B1 |
8754757 | Ullrich et al. | Jun 2014 | B1 |
8754758 | Ullrich et al. | Jun 2014 | B1 |
9164587 | Da Costa | Oct 2015 | B2 |
9171437 | Nakamura et al. | Oct 2015 | B2 |
9174134 | Grant | Nov 2015 | B1 |
9367136 | Latta et al. | Jun 2016 | B2 |
9370459 | Mahoney | Jun 2016 | B2 |
9370704 | Marty | Jun 2016 | B2 |
9392094 | Hunt et al. | Jul 2016 | B2 |
9462262 | Worley, III et al. | Oct 2016 | B1 |
9514620 | Bhatia | Dec 2016 | B2 |
9519346 | Lacroix | Dec 2016 | B2 |
9557830 | Grant | Jan 2017 | B2 |
9619029 | Lacroix | Apr 2017 | B2 |
9626805 | Lampotang et al. | Apr 2017 | B2 |
9645646 | Cowley et al. | May 2017 | B2 |
9652037 | Rubin et al. | May 2017 | B2 |
9753542 | Chizeck | Sep 2017 | B2 |
9760166 | Ammi et al. | Sep 2017 | B2 |
9811854 | Lucido | Nov 2017 | B2 |
9814974 | Grant | Nov 2017 | B2 |
9849379 | Alghooneh | Dec 2017 | B2 |
9851799 | Keller et al. | Dec 2017 | B2 |
9852590 | Bhatia | Dec 2017 | B2 |
9933851 | Goslin et al. | Apr 2018 | B2 |
9948885 | Kurzweil | Apr 2018 | B2 |
10185396 | Rihn | Jan 2019 | B2 |
10226697 | Schmitz | Mar 2019 | B2 |
10254838 | Venkatesan | Apr 2019 | B2 |
20020024503 | Armstrong | Feb 2002 | A1 |
20020080112 | Braun | Jun 2002 | A1 |
20030030619 | Martin et al. | Feb 2003 | A1 |
20040056840 | Goldenberg et al. | Mar 2004 | A1 |
20050134561 | Tierling | Jun 2005 | A1 |
20050151720 | Cruz-Hernandez et al. | Jul 2005 | A1 |
20060109256 | Grant et al. | May 2006 | A1 |
20060290662 | Houston | Dec 2006 | A1 |
20070013677 | Rosenberg et al. | Jan 2007 | A1 |
20070182708 | Poupyrev et al. | Aug 2007 | A1 |
20070265077 | Tom | Nov 2007 | A1 |
20070290988 | Nogami | Dec 2007 | A1 |
20080223627 | Lacroix et al. | Sep 2008 | A1 |
20080294984 | Ramsay et al. | Nov 2008 | A1 |
20090088659 | Graham et al. | Apr 2009 | A1 |
20090122006 | Nielsen | May 2009 | A1 |
20090033624 | Rosenberg et al. | Dec 2009 | A1 |
20090297328 | Slocum, III | Dec 2009 | A1 |
20100122006 | Kanaya | May 2010 | A1 |
20100148943 | Rosenberg et al. | Jun 2010 | A1 |
20100151948 | Vance | Jun 2010 | A1 |
20100173686 | Grant | Jul 2010 | A1 |
20100231541 | Cruz-Hernandez et al. | Sep 2010 | A1 |
20110006888 | Bae et al. | Jan 2011 | A1 |
20110018697 | Birnbaum | Jan 2011 | A1 |
20110163860 | Ryu | Jul 2011 | A1 |
20110202155 | Ullrich et al. | Aug 2011 | A1 |
20110215913 | Ullrich et al. | Sep 2011 | A1 |
20110244963 | Grant | Oct 2011 | A1 |
20110264491 | Birnbaum et al. | Oct 2011 | A1 |
20120068970 | Pemberton-Pigott | Mar 2012 | A1 |
20120206246 | Cruz-Hernandez et al. | Aug 2012 | A1 |
20120206247 | Bhatia et al. | Aug 2012 | A1 |
20120306631 | Hughes | Dec 2012 | A1 |
20130131851 | Ullrich et al. | May 2013 | A1 |
20130147610 | Grant et al. | Jun 2013 | A1 |
20130194085 | Grant et al. | Aug 2013 | A1 |
20130207792 | Lim et al. | Aug 2013 | A1 |
20130207917 | Cruz-Hernandez et al. | Aug 2013 | A1 |
20130222311 | Pesonen | Aug 2013 | A1 |
20130265286 | Da Costa et al. | Oct 2013 | A1 |
20130267321 | Burgess et al. | Oct 2013 | A1 |
20130324254 | Huang et al. | Dec 2013 | A1 |
20130335209 | Cruz-Hernandez et al. | Dec 2013 | A1 |
20140002386 | Rosenberg et al. | Jan 2014 | A1 |
20140064516 | Cruz-Hernandez et al. | Mar 2014 | A1 |
20140118125 | Bhatia | May 2014 | A1 |
20140315642 | Grant | Oct 2014 | A1 |
20140366635 | Zusman | Dec 2014 | A1 |
20150070146 | Cruz-Hernandez | Mar 2015 | A1 |
20150070147 | Cruz-Hernandez | Mar 2015 | A1 |
20150070148 | Cruz-Hernandez | Mar 2015 | A1 |
20150070260 | Saboune | Mar 2015 | A1 |
20150070261 | Saboune | Mar 2015 | A1 |
20150070265 | Cruz-Hernandez | Mar 2015 | A1 |
20150130706 | Lacroix | May 2015 | A1 |
20150130707 | Da Costa | May 2015 | A1 |
20150352437 | Koseki | Dec 2015 | A1 |
20160070348 | Cowley et al. | Mar 2016 | A1 |
20160084605 | Monti | Mar 2016 | A1 |
20160086457 | Baron et al. | Mar 2016 | A1 |
20160132114 | Rihn | May 2016 | A1 |
20160132116 | Grant | May 2016 | A1 |
20160163227 | Penake et al. | Jun 2016 | A1 |
20160166930 | Bray et al. | Jun 2016 | A1 |
20160169635 | Hannigan et al. | Jun 2016 | A1 |
20160170508 | Moore et al. | Jun 2016 | A1 |
20160171860 | Hannigan et al. | Jun 2016 | A1 |
20160171908 | Moore et al. | Jun 2016 | A1 |
20160175703 | Venkatesan | Jun 2016 | A1 |
20160179220 | Venkatesan | Jun 2016 | A1 |
20160187969 | Larsen et al. | Jun 2016 | A1 |
20160187974 | Mallinson | Jun 2016 | A1 |
20160201888 | Ackley et al. | Jul 2016 | A1 |
20160209658 | Zalewski | Jul 2016 | A1 |
20160214015 | Osman et al. | Jul 2016 | A1 |
20160214016 | Stafford | Jul 2016 | A1 |
20160375170 | Kursula et al. | Dec 2016 | A1 |
20170102771 | Lei | Apr 2017 | A1 |
20170103574 | Faaborg et al. | Apr 2017 | A1 |
20170131775 | Clements | May 2017 | A1 |
20170148281 | Do et al. | May 2017 | A1 |
20170154505 | Kim | Jun 2017 | A1 |
20170168576 | Keller et al. | Jun 2017 | A1 |
20170168773 | Keller et al. | Jun 2017 | A1 |
20170178407 | Gaidar et al. | Jun 2017 | A1 |
20170203207 | Lacroix | Jul 2017 | A1 |
20170203221 | Goslin et al. | Jul 2017 | A1 |
20170203225 | Goslin | Jul 2017 | A1 |
20170206709 | Goslin et al. | Jul 2017 | A1 |
20170212591 | Churikov | Jul 2017 | A1 |
20170214782 | Brinda | Jul 2017 | A1 |
20170257270 | Goslin et al. | Sep 2017 | A1 |
20170333788 | Grant | Nov 2017 | A1 |
20170352185 | Bonilla Acevedo et al. | Dec 2017 | A1 |
20180050267 | Jones | Feb 2018 | A1 |
20180053351 | Anderson | Feb 2018 | A1 |
20180074589 | Da Costa | Mar 2018 | A1 |
20180077976 | Keller et al. | Mar 2018 | A1 |
20180081436 | Keller et al. | Mar 2018 | A1 |
20180093181 | Goslin et al. | Apr 2018 | A1 |
20180107277 | Keller et al. | Apr 2018 | A1 |
20180120936 | Keller et al. | May 2018 | A1 |
20180188812 | Grant | Jul 2018 | A1 |
Number | Date | Country |
---|---|---|
1983125 | Jun 2007 | CN |
102822772 | Dec 2012 | CN |
102906667 | Jan 2013 | CN |
103365415 | Oct 2013 | CN |
104107539 | Oct 2014 | CN |
0864144 | Sep 1998 | EP |
0875819 | Nov 1998 | EP |
2039404 | Mar 2009 | EP |
2648071 | Oct 2013 | EP |
2796965 | Oct 2014 | EP |
2001-137553 | May 2001 | JP |
2003199974 | Jul 2003 | JP |
2003-534620 | Nov 2003 | JP |
2005-056267 | Mar 2005 | JP |
2008-173186 | Jul 2008 | JP |
2009-233108 | Oct 2009 | JP |
2009-535694 | Oct 2009 | JP |
2010-165383 | Jul 2010 | JP |
2013-524351 | Jun 2013 | JP |
2013-164845 | Aug 2013 | JP |
2013218327 | Oct 2013 | JP |
2014-180572 | Sep 2014 | JP |
2013-0040853 | Apr 2013 | KR |
2003032289 | Apr 2003 | WO |
2010065207 | Jun 2010 | WO |
2013116247 | Aug 2013 | WO |
Entry |
---|
Conan Hughes, “Turn Off Wii Remote's Rumble Feature”, Apr. 9, 2011, XP055172295, http://www.conanhughes.com/2011/04/turn-off-wii-remotes-rumble-feature.html. |
Mkwone, “Move FAQ—Playstation Forum”, Sep. 3, 2010, XP055172297, http://community.eu.playstation.com/t5/PlayStation-Move/Move-FAQ/td-p/11447251. |
Number | Date | Country | |
---|---|---|---|
20170203207 A1 | Jul 2017 | US |
Number | Date | Country | |
---|---|---|---|
61904342 | Nov 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14539111 | Nov 2014 | US |
Child | 15454286 | US |