The present disclosure belongs to the technical field of emotion regulation, and particularly relates to an affective haptic regulation system and method based on multimodal fusion.
In recent years, the rapid development of affective computing and haptic technology has given rise to an emerging field: affective haptics. Affective computing aims to reveal a mechanism of emotion generation and expression, while haptic technology focuses on simulating human haptic sensing. Integrating affective information with haptic technology, affective haptics aims to explore the potential of using haptic technology in affective detection, display, and communication, providing new possibilities for human-computer interaction.
As core technology in the field of affective haptics, an affective haptic regulation system aims to perceive an affective state of an individual in real time, and guide it through haptic stimuli, thereby achieving proactive affective regulation during an interaction process. The affective haptic regulation system has promising applications in the fields of human-computer interaction, such as medical rehabilitation and audio-visual entertainment. For example, in the medical field, the affective haptic regulation system provides innovative methods for affective disorders such as depression by assisting in affective regulation through haptic stimuli. In the field of audio-visual entertainment, the affective haptic regulation system enhances affective immersion through haptic stimuli, creating a more immersive experience for a user.
However, despite broad prospects of the affective haptic regulation system, there are still some technical difficulties that are imperative to be solved:
The foregoing two technical difficulties restrict further development of the affective haptic regulation system.
In order to solve the above problems, the present disclosure discloses an affective haptic regulation system and method based on multimodal fusion, the system fuses multi-physiological signal features with audio and haptic modal features by acquiring a plurality of physiological signals of a user, such as EEG and ECG signals, and can accurately identify a current affective state of the user in real time in combination with advanced data processing and analysis technology, seek a haptic parameter with the help of an optimization theory, and achieve proactive regulation of affective state by applying haptic stimuli to the user.
In order to achieve the above objective, the present disclosure adopts a following technical solution:
an affective haptic regulation system based on multimodal fusion, including a haptic optimal parameter adjustment module, a haptic generation module, a visual-auditory generation module, a multi-physiological signal acquisition module, a multi-sensory signal acquisition module, and a multimodal fusion emotion recognition module. The haptic optimal parameter adjustment module automatically solves an optimal haptic parameter according to a difference between current emotion and target emotion of the user, and sends the optimal haptic parameter to the haptic generation module and generates a haptic effect; the haptic generation module and the visual-auditory generation module cooperate to generate visual-auditory-haptic fusion stimuli to act on the user, so as to regulate and control emotions of the user; the multi-physiological signal acquisition module and the multi-sensory signal acquisition module acquire various physiological signals, audio signals and haptic vibration signals of the user in real time, and the signals are inputted to the multimodal fusion emotion recognition module to detect a current affective state of the user, which is then sent back to the haptic optimal parameter adjustment module to form a closed-loop affective haptic regulation system.
As a core of the affective haptic regulation system based on multimodal fusion, the haptic optimal parameter adjustment module automatically seeks for a haptic parameter with the help of an optimization theory according to the difference between the current affective state and the target emotion of the user, and sends the haptic parameter to the haptic generation module to ensure the effectiveness of affective regulation. The haptic optimal parameter adjustment module includes a haptic parameter optimization model and a haptic parameter solving module. The haptic parameter optimization model is expressed as:
in the formulae, Mi is an actual power value of a certain electrode i on a brain topographic map, Mbi is a reference power value of the certain electrode i on the brain topographic map, Sis a calculated affective state value, Sb is a target affective state value, P is an actual power value consumed by the haptic generation module, Pm is a set maximum power value consumed by the haptic generation module, fm is a set maximum haptic vibration frequency, qm is a set maximum haptic vibration intensity, rm is a set maximum haptic vibration rhythm, and γ, μ and φ are weight coefficients of the haptic parameter optimization model.
The haptic parameter solving module adopts a particle swarm optimization algorithm, reinforcement learning, and other machine learning algorithms, to solve four parameters in the haptic parameter optimization model, that is, a haptic vibration frequency f, a haptic vibration intensity q, a haptic vibration rhythm r, and a haptic vibration position c, and to send the solved parameters to the haptic generation module.
The haptic generation module is a wearable device capable of expressing haptic sensation through vibration, including a vibration vest, a vibration bracelet, vibration gloves, and the like, and the haptic generation module is configured to transmit specific haptic experience to the user by setting a vibration frequency, a vibration intensity, a vibration rhythm and a vibration position of a haptic generation device. The haptic generation module exhibits continuous presence of a background haptic vibration that changes adaptively with audio, and at the same time, another haptic expression can be realized based on the four parameters calculated by the haptic parameter optimization model, and the haptic expression corporates with the background haptic vibration to effectively enhance affective experience of the user.
The visual-auditory generation module provides the user with visual and auditory stimuli, including movie clips of different emotion types. The movie clips are conducive to guiding the user to enter a specific affective state, where audio of the movie clips provides a basis for changes in the background haptic vibration.
The multi-physiological signal acquisition module acquires a plurality of physiological signals of the user in real time, including 64-channel electroencephalogram (EEG) signals and electrocardiogram (ECG) signals, where the EEG signals are acquired by an EEG signal acquisition module, and the EEG signal acquisition module is composed of a 64-channel actiCAP electrode cap and an EEG amplifier from Brain Products GmbH. The ECG signals are acquired by an ECG signal acquisition module, which is an ActiveTwo series high-channel ECG acquisition system from Biosemi B. V.
The multi-sensory signal acquisition module includes an auditory signal acquisition module and a haptic signal acquisition module, which are capable of acquiring audio signals generated in the visual-auditory generation module in real time, and acquiring the haptic vibration signals generated in the haptic generation module in real time, respectively.
The multimodal fusion emotion recognition module is capable of analyzing, processing and recognizing the current affective state of the user according to multi-physiological signals of the user and multi-sensory signals of emotion elicitation materials, and then sending a signal of the current affective state to the haptic optimal parameter adjustment module for intelligent regulation of a haptic parameter. The multimodal fusion emotion recognition module includes a signal preprocessing module, a feature extraction module, a feature fusion module, and an emotion decoding module. The signal preprocessing module pre-processes the acquired EEG signals and the ECG signals, including downsampling, filtering, artifact removal, and the like. The feature extraction module extracts features from the audio signals, the haptic vibration signals, the preprocessed EEG signals and ECG signals, respectively. The feature fusion module performs feature fusion on the multi-physiological signal features, audio features extracted from the audio signals and vibration features extracted from the haptic vibration signals by using a feature fusion algorithm. The emotion decoding module classifies the multimodal fusion features by using a classification algorithm to obtain the current affective state of the user. The multimodal fusion emotion recognition module has the advantage of fusing the multi-physiological signal features with audio and haptic modal features, which can effectively reduce the influence of unstable physiological signals on emotion recognition results.
The present disclosure has the beneficial effects:
The present disclosure will be further illustrated below with reference to the accompanying drawings and specific embodiments. It should be understood that the following specific embodiments are only used to illustrate the present disclosure, but are not intended to limit the scope of the present disclosure.
This embodiment describes an affective haptic regulation system based on multimodal fusion, with an overall schematic diagram shown in
Referring to
The haptic parameter optimization model 7 establishes a reliable mathematical model based on elements such as an electroencephalogram signal, an affective state, and vibration consumption power, and is configured to adaptively adjust four key parameters in the haptic generation module 2, including a haptic vibration frequency f, a haptic vibration intensity q, a haptic vibration rhythm r, and a haptic vibration position c. The haptic parameter optimization model 7 is expressed as:
in the formulae, Mi is an actual power value of a certain electrode i on a brain topographic map, Mbi is a reference power value of the certain electrode i on the brain topographic map, Sis a calculated affective state value, Sb is a target affective state value, P is an actual power value consumed by the haptic generation module, Pm is a set maximum power value consumed by the haptic generation module, fm is a set maximum haptic vibration frequency, qm is a set maximum haptic vibration intensity, rm is a set maximum haptic vibration rhythm, and γ, μ and φ are weight coefficients of the haptic parameter optimization model 7.
The haptic parameter solving module 8 adopts an optimization algorithm, machine learning, reinforcement learning, and the like, and is configured to solve the four parameters in the haptic parameter optimization model 7, that is, the haptic vibration frequency f, the haptic vibration intensity q, the haptic vibration rhythm r, and the haptic vibration position c, and to send the solved parameters to the haptic generation module 2.
Referring to
The visual-auditory generation module 3 provides the user with visual and auditory stimuli, including movie clips of different emotion types, and guides the user to enter a specific affective state. Specifically, visual-auditory material includes 16 movie clips of about 4 minutes, covering four types of emotion: happiness, sadness, fear and calmness, that is, each of the types of emotion corresponds to 4 movie clips. Audio of the movie clips provides a basis for changes in the background haptic vibration for the haptic generation module 2, and the visual-auditory stimuli cooperate with the haptic stimuli to further enhance the affective experience of the user.
Referring to
Referring to
Referring to
This embodiment provides a technical solution: an affective haptic regulation method based on multimodal fusion. Referring to
Step S1: starting an experiment, and applying visual-auditory-haptic fusion stimuli. The user is guided into different affective states by presenting visual-auditory stimuli and haptic stimuli. Referring to
Step S2: performing multimodal acquisition, including multi-physiological signals and multi-sensory signals. EEG signals, ECG signals, the audio signals, and haptic vibration signals of the user are acquired in real time. The EEG signals are acquired by the EEG signal acquisition module composed of the 64-channel actiCAP electrode cap and the EEG amplifier from Brain Products GmbH; the ECG signals are acquired by the ECG signal acquisition module composed of the ActiveTwo series high-channel ECG acquisition system from Biosemi B.V.; the audio signals are acquired by the auditory signal acquisition module, and the haptic vibration signals are acquired by the haptic signal acquisition module.
Step S3: performing multimodal feature extraction and feature fusion. The acquired EEG signals and the ECG signals are preprocessed, including downsampling, filtering, artifact removal, and the like, to ensure the quality and stability of the signals. Features are then extracted from the preprocessed EEG and ECG signals to obtain EEG and ECG signal features of the user, and these features can capture variation patterns of different physiological signals in different affective states. Furthermore, the audio features are extracted from the audio signals, and the haptic vibration features are extracted from the haptic vibration signals. The feature fusion algorithm is adopted to fuse the multi-physiological signal features with the audio features and the haptic vibration features to enhance the accuracy and robustness of emotion recognition.
Step S4: decoding multimodal fusion emotion and giving feedback. The multimodal fusion features are classified by using the classification algorithm to obtain the current affective state of the user, and the current affective state is then sent back to the haptic optimal parameter adjustment module.
Step S5: solving and updating the vibration parameters. The haptic optimal parameter adjustment module receives the current affective state of the user, automatically solves the optimal haptic parameter according to the difference between the current emotion and the target emotion, and sends the optimal haptic parameter to the haptic generation module and generates the haptic effect, so as to ensure that the applied haptic stimuli match an actual affective need of the user.
Step S6: finishing the experiment and establishing an affective haptic database. Completion of the playing of the 16 movie clips indicates that the entire experiment is finished, and the haptic the vibration parameters corresponding to different affective states of the user are the analyzed, and the affective haptic database of the user is then established. Different haptic vibration parameters are mapped to different affective states, and a personalized haptic mode is generated to present diverse affective experience for the user.
It should be noted that the above content merely illustrates the technical idea of the present disclosure and cannot limit the protection scope of the present disclosure, those of ordinary skill in the art may also make some modifications and improvements without departing from the principle of the present disclosure, and these modifications and improvements should also fall within the protection scope of the claims of the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
202311121644.1 | Sep 2023 | CN | national |
This application is a continuation of PCT application No. PCT/CN2023/127190, filed on Oct. 27, 2023, which claims a priority benefit to China patent application No. CN202311121644.1 filed on Sep. 1, 2023. The entirety of the above-mentioned patent applications are hereby incorporated by reference herein and made a part of this specification.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2023/127190 | Oct 2023 | WO |
Child | 18817208 | US |