This patent application claims the benefit and priority of Chinese Patent Application No. 202410087933.2 filed with the China National Intellectual Property Administration on Jan. 22, 2024, the disclosure of which is incorporated by reference herein in its entirety as part of the present application.
The present disclosure belongs to the technical field of biometric identification based on electroencephalogram signals, and in particular relates to a cross-period brain fingerprint identification method with paradigm adaptive decoupling and a system thereof.
As a highly identifiable physiological feature, an electroencephalogram signal has great potential in the field of biometric identification. With the continuous development and popularization of consumer wearable devices, users can acquire electroencephalogram data through sensors in the devices to realize contactless identity identification. Compared with traditional biometric technologies (such as face identification and fingerprint identification), the electroencephalogram signal identification is more confidential, more difficult to be stolen, and revocable and revisable. On the other hand, compared with conventional biological features, the electroencephalogram signal identification has obvious advantages in some specific scenarios, such as some highly confidential and safe occasions, or in some new technology usage scenarios, such as virtual reality (VR) and augmented reality (AR).
However, different methods of collecting electroencephalogram signals lead to different features and manifestations of electroencephalogram signals, which challenges the generalization and identification ability of the model. In addition, electroencephalogram signals are very sensitive to internal and external noises, such as physiological state, psychological state and environmental noise, which makes it difficult to guarantee the stability and reliability of the model in different periods. Therefore, most of the work is to collect electroencephalogram signals in a single paradigm or in a single period, which is not consistent with the actual application scenario.
In order to overcome these challenges, the present disclosure provides a cross-period brain fingerprint identification method with paradigm adaptive decoupling. The method can effectively separate identity-related features and paradigm task-related features from highly coupled electroencephalogram information; and further learn domain invariant features with identity identification ability through domain adversarial training. The present disclosure can be adaptive to electroencephalogram signals collected by various paradigms under a cross-period condition, which has a high practical value in practical application scenarios and provides innovative solutions for the fields of biometric identification, identity verification and the like.
According to the shortcomings of the prior art, the present disclosure provides a cross-period brain fingerprint identification method with paradigm adaptive decoupling and a system thereof. First, a feature extractor extracts is a feature representation from original electroencephalogram data; thereafter, effectively separates identity-related features and paradigm task-related features from highly coupled electroencephalogram information; and finally, learns domain invariant features with identity identification ability through domain adversarial training.
In a first aspect, the present disclosure provides a cross-period brain fingerprint identification method with paradigm adaptive decoupling, including the following steps:
Step 4:, constructing a decoupling module for decoupling features and respective classifiers, and training and testing the decoupling module and the classifiers;
Preferably, in Step 2, preprocessing the electroencephalogram data includes: filtering and down-sampling the electroencephalogram data collected in Step 1, and then fragmenting the electroencephalogram data to obtain a plurality of fragments with a sample length of L.
Preferably, in Step 2, dividing the electroencephalogram data into the source domain and the target domain according to the sequence of collecting periods means that: in a time sequence, the period data collected first is taken as the source domain with identity labels which is denoted as S={XS, YS}={(xS1,yS1), ⋅ ⋅ ⋅ , (xSn
d×L represents an electroencephalogram data fragment in a period of time, ysi|i=1n
C represents an identity label of an i-th subject in the source domain, and C represents the number of the subjects; the period data collected later is taken as unlabeled target domain of identities to be predicted, which is denoted as
T={XT}={xT1, ⋅ ⋅ ⋅ , xTn
d×L represents an electroencephalogram data fragment in a period of time, and nt represents the number of samples in the target domain.
Preferably, in Step 3, the kernel size of each type of one-dimensional convolution layer is determined by the proportion coefficient ak|k=1K∈ and the sample length L, and the proportion coefficient is artificially set, where k represents the convolution kernel of a k-th type of one-dimensional convolution layers.
In the present disclosure, ak=[0.1, 0.2, 0.5], that is, there are three types of one-dimensional convolution kernels with different sizes.
The scale of the time convolution kernel of the k-th type is denoted as k and defined as:
Preferably, in Step 3, the multi-scale one-dimensional convolution layer is specifically expressed as:
wherein Xi∈C×L˜
S/T,i∈[1, ⋅ ⋅ ⋅ ,N] represents the input of the multi-scale one-dimensional convolution layer, that is, the preprocessed electroencephalogram data, N is the number of sample fragments of the preprocessed electroencephalogram data, C is the number of channels of the electroencephalogram collecting device, and L is the length of the set electroencephalogram data in the time dimension; XT_outk∈
t×C×f
Conv2D(Xi,STk) represents the convolution operation of Xi using a time convolution kernel with a size of STk; Φsquare(·) is a square function;
AP(·) represents the average pooling operation, and Φlog(·) represents a logarithmic function.
The splicing layer is specifically expressed as:
where Γ(·) represents the serial operation along the feature dimension; Xcati∈t×C×Σf
The fusion layer uses a 1*1 convolution layer to fuse the features learned by different convolution kernels; the number of convolution kernels in the 1*1 convolution layer is set to t. Preferably, Leaky-ReLU is used as the activation function, and the average pooling layer is used to down-sample the learned representation. After batch standardization, the fused representation from different 1*1 convolution kernels will be flattened and become the attribute of each channel node in the graph representation; therefore, the attention fusion representation
where bn(·) is a batch normalization function,
dropout(·) is a random dropout layer preventing over-fitting,
fuse(·) is a 1*1 convolution function, ΦL-ReLU(·) is a Leaky-ReLU activation function;
AP(·) represents the average pooling operation.
i
fuse is reshaped as Xifuse∈C×t* 0.5*Σf
The graph convolution module includes a plurality of graph convolution layers connected in series. Specifically, a messaging framework is used to realize the graph convolution layer (GNN):
where A=XifuseXifuseT∈d×d, d represents the number of channels of electroencephalogram features;
The attention embedding module obtains the embedded vector representation of the features through the attention mechanism. Specifically, first, the global mean representation
is calculated, and then each channel xouti|id and hmean are subjected to inner products to obtain a similarity representation, which can be regarded as the expression of the importance degree belonging to each channel. Finally, the final embedded vector representation is obtained by weighted summation:
Preferably, in the training process, the decouplers use adversarial training to achieve feature decoupling, specifically, the intra-domain specific identity information decoupler, the inter-domain invariant identity information decoupler and the paradigm task information decoupler decouple hout into the domain specific identity feature representation hsped-id, the domain invariant identity feature representation hinv-id and the paradigm task-related feature representation htask, respectively.
Preferably, the intra-domain specific identity feature hsped-id and the inter-domain invariant identity feature hinv-id are constrained by L1 norm.
Thereafter, the identity classifier c-id(·) and the paradigm task classifier
c-task are trained to realize correct classification, which are iteratively optimized through a cross entropy loss.
where (x
x˜{tilde over (D)}
In the training process, the domain classifier C-domain(·) realizes the distribution alignment of the source domain and the target domain in a domain adversarial manner, specifically, in an identity-related feature subspace, the source domain and the target domain are aligned through domain adversarial training:
where x˜{tilde over (D)}
C-domain(hinv-id)] represents that an encouragement that the domain classifier
C-domain(·) correctly predicts the source domain data, and
x˜{tilde over (D)}
C-domain(hinv-id)] represents an encouragement that the features decoupled by the inter-domain invariant identity information decoupler deceive
C-domain(·) to obtain the domain invariant identity feature.
Preferably, a better decoupling effect is obtained by updating the network through the mutual information loss. That is, the first mutual information loss 1(hinv-id,htask) between the domain invariant identity feature representation hinv-id and the paradigm task-related feature htask is calculated through the first mutual information network, and the second mutual information loss
2 (hinv-id, hspec-id) between the domain invariant identity feature representation hinv-id and the domain specific identity feature representation hsped-id is calculated through the second mutual information network, which are expressed as:
where htask and hspec-id represent the edge distributions sampled from htask and hsped-id, respectively; 1/2 (·) represents the first mutual information network and the second mutual information network, respectively; θ is a learnable parameter.
In a second aspect, the present disclosure provides a cross-period brain fingerprint identification system, including:
In a third aspect, the present disclosure provides a computer-readable storage medium having a computer program stored thereon, which, when executed in a computer, causes the computer to execute the method.
In a fourth aspect, the present disclosure provides a computing device, including a memory and a processor, wherein executable codes are stored in the memory, and the processor, when executing the executable codes, implement the method.
Through these devices and apparatuses, the present disclosure realizes an efficient and accurate cross-period brain fingerprint identification solution, and brings an innovative technical support for the fields of biometric identification, identity verification and the like.
The present disclosure has the following features and beneficial effects.
In order to more clearly explain the technical scheme in the embodiment of the present disclosure or the prior art, the drawing required in the description of the embodiment or the prior art will be briefly introduced hereinafter. Obviously, the drawing in the following description only represents some embodiments of the present disclosure. For those skilled in the art, other related drawings can be obtained according to the drawing without creative labor.
In the case of no contradiction, the embodiments in the present disclosure and the features in the embodiments can be combined with each other. The specific implementation steps of the cross-period brain fingerprint identification method with paradigm adaptive decoupling of the present disclosure are as follows.
3. A feature extraction module with paradigm adaptive decoupling is constructed: the feature extraction module with paradigm adaptive decoupling mainly consists of a multi-scale convolution module, a graph convolution module and an attention embedding module. The multi-scale convolution module includes more than one one-dimensional convolution layer, which has convolution kernels with different sizes. The multi-scale convolution module further includes a splicing layer and a fusion layer. The graph convolution module consists of a plurality of parallel graph convolution networks. The attention embedding module acts on an output of graph convolution of different levels, and transforms a graph structure into an embedding vector through an attention mechanism.
4. A decoupling module for decoupling features and respective classifiers thereof are constructed: wherein the decoupling module consists of a plurality of decouplers and mutual information networks. The decouplers includes an intra-domain specific identity information decoupler, an inter-domain invariant identity information decoupler and a paradigm task information decoupler. The decouplers decouple extracted feature representation into domain specific identity feature representation, domain invariant identity feature representation and paradigm task-related feature representation. Both the decouplers and the mutual information networks all consist of fully connected layer networks, and outputs of the decouplers are used as inputs of the classifiers. The mutual information networks aim at reducing mutual information between features and promote better decoupling. The classifiers includes: a domain classifier, which is configured to perform domain adversarial training to reduce distribution differences between different domains and obtain domain invariant features; an identity classifier, which aims at obtaining accurate classification information; a paradigm task classifier, which aims at promoting decoupling effect and reducing influence of interference information on an identity identification task. Each classifier takes the output of the respective decoupler as an input, which consists of a fully connected layer and a Softmax activation function.
5. The network model is trained: the classifiers are trained by using decoupled domain invariant identity information, and model parameters are constantly updated through an optimization algorithm, so that the model can obtain better identification performance on training data.
Through above specific implementation steps, the present disclosure realizes a cross-period brain fingerprint identification method with paradigm adaptive decoupling, which can effectively improve the accuracy and robustness of brain fingerprint identification in practical application scenarios.
Specifically, in Step 1, electroencephalogram cap leads are connected to corresponding brain region of a subject to collect electroencephalogram data.
In Step 2, the electroencephalogram data collected in Step 1 is preprocessed. Original electroencephalogram signal contains frequencies of noises. In order to remove power frequency interference resulted from electroencephalogram collecting device and electromyogram interference of the subject, the electroencephalogram data is down-sampled to 200 Hz, and the original electroencephalogram data is filtered by a Butterworth filter at 1 to 75 Hz.
Obtained electroencephalogram features are divided into fragments with a predetermined time window size of L. The specific window size of the present disclosure is 5 s, and the corresponding fragments are labeled with a label of a subject. The fragments are divided into a source domain and a target domain according to the sequence of collecting periods. The period data collected first is taken as the source domain with identity labels, and the period data collected later is taken as the target domain of the identity labels to be predicted.
According to the further setting of the present disclosure, in Step 3, a feature extraction module with paradigm adaptive decoupling is included. Specifically, the feature extraction module mainly consists of a multi-scale convolution module, a graph convolution module and an attention embedding module. The multi-scale convolution module includes more than one one-dimensional convolution layer, which has convolution kernels with different sizes. The multi-scale convolution module further includes a splicing layer, a fusion layer and a transition layer. A size of a one-dimensional convolution kernel is determined according to different proportions of the sample duration L, and the proportion coefficient is denoted as ak|k=1K∈, where k represents a convolution kernel of a k-th type of time convolution layer. In the present disclosure, ak=[0.1, 0.2, 0.5], that is, there are three types of one-dimensional convolution kernels with different sizes. The scale of the k-th type of the time convolution kernel is denoted as
k and defined as:
For a given electroencephalogram data xi|i=1n∈d×L˜
S/T, n is the number of samples of the electroencephalogram data, and d is the number of channels for collecting the electroencephalogram data. First, a plurality of multi-scale one-dimensional time convolution kernels will be used at the same time to learn dynamic time representation of the electroencephalogram data. The representation obtained after the multi-scale time convolution layer is processed by the logarithm of the average pooling square in order to learn the power features of the dynamic time representation of the electroencephalogram signals.
The time convolution output of the k-th type is denoted as X(i,k)∈T
where Conv2D(·) represents the convolution operation, Φsquare(·) is a square function;
AP(·) represents the average pooling operation, and Φlog(·) represents a logarithmic function.
The splicing layer will connect the time convolution kernel outputs of all levels in series in the feature dimension. Therefore, for the input xi|i=1n∈d×L, the output Xcati∈
T
where Γ(·) represents the serial operation along the feature dimension.
Preferably, the fusion layer uses a 1*1 convolution layer as the attention fusion layer to fuse the features learned by different convolution kernels. The number of convolution kernels in the 1*1 convolution layer is set to t. Preferably, Leaky-ReLU is used as the activation function, and the average pooling layer is used to down-sample the learned representation. After batch standardization, the fused representation from different 1*1 convolution kernels will be flattened and become the attribute of each channel node in the graph representation. Therefore, the attention fusion representation
where bn(·) is a batch normalization function,
dropout(·) is a random dropout layer preventing over-fitting,
fuse(·) is a 1*1 convolution function, ΦL-ReLU(·) is a Leaky-ReLU activation function.
t×C×0.5*Σf
C×t*0.5*Σf
The graph convolution module is realized by a graph convolution network using a general messaging framework:
where A=XifuseXifuseT∈d×d
The attention embedding module obtains the importance of each channel through a similarity calculation, and then carries out a weighted summation based on this. Specifically, first, the global mean representation is calculated:
Thereafter, each channel xouti|i d and hmean are subjected to inner product to obtain a similarity representation, which can be regarded as the expression of the importance degree belonging to each channel. Finally, the final embedded vector representation is obtained by weighted summation:
According to the further setting of the present disclosure, in Step 4, a decoupling module for decoupling features and respective classifiers thereof are constructed, and are trained and tested.
Specifically, the adversarial training method is used to realize feature decoupling and classification. First, hout is decoupled into an identity-related feature had and a task-related feature htask by a decoupler D(·), and then the classifier
C-id(·) is trained to realize correct classification, which is iteratively optimized through a cross entropy loss.
Thereafter, the classifiers are fixed, and the two decoupling features of the decouplers are trained to deceive the classifiers of each other. The present disclosure uses a minimized negative entropy to encourage such deception effect:
The first item indicates deceiving the identity classifier with task features, and the second item indicates deceiving the task classifier with identity features.
The present disclosure realizes the distribution alignment of the source domain and the target domain through adversarial training, and specifically, the training loss function can be expressed as:
Preferably, the present disclosure further decouples the identity information into intra-domain specific identity information and inter-domain invariant identity information, and reduces the difference of prediction results therebetween through L1 norm constraint, specifically, which can be expressed as:
Preferably, a better decoupling effect is obtained by updating the network through the mutual information loss. That is, the mutual information loss between the inter-domain invariant identity feature hinv-id and the paradigm task-related feature htask is calculated through the first mutual information network, and the mutual information loss between the inter-domain invariant identity feature hinv-id and the intra-domain specific identity feature hsped-id is calculated through the second mutual information network, which are specifically expressed as:
where htask and hspec-id represent the edge distributions sampled from htask and hsped-id, respectively; 1/2(·) represent the first mutual information network and the second mutual information network, respectively, that is, the fully connected layer network module; and θ is a learnable parameter.
According to the further setting of the present disclosure, in Step 5, the network model is trained. In order to optimize the network parameters, the present disclosure uses a back propagation method, and updates the network parameters through iteration until the required standard is reached. The training update mode is as shown in the following table 1.
S , unlabeled
T , feature
The cross-period brain fingerprint identification method with paradigm adaptive decoupling according to the present disclosure is compared with the classic and recently proposed methods in the field of a brain-computer interface on two cross-period data sets: SEED-V and an electroencephalogram data set collected based on an RSVP protocol under the domain adaptive framework, and the obtained identification precisions (%) are as shown in the following table:
The embodiments of the present disclosure have been described in detail above with reference to the drawing, but the present disclosure is not limited to the described embodiments. It will be obvious to those skilled in the art that many changes, modifications, substitutions and variations can be made to these embodiments, including components, without departing from the principle and spirit of the present disclosure, which still fall within the scope of protection of the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
202410087933.2 | Jan 2024 | CN | national |