The present application relates to the technical field of machine learning of face image processing, in particular to a face image clustering method and system based on a localized simple multiple kernel k-means.
With the widespread application of the face recognition and retrieval system, the face image data in the system has increased sharply, and the face clustering technology has become an important foundation for improving the retrieval efficiency of the system. In face clustering, the face image information in the database is usually clustered into different subclasses, so that the similarity between the subclasses is as small as possible, and the similarity in the subclasses is as large as possible. In this way, during retrieval, only the subclasses with high similarity to the target to be retrieved need to be identified one by one, thus retrieving records with the highest similarity to the target to be retrieved.
K-means clustering is the most widely used method, and kernel k-means clustering is widely studied because it can learn the nonlinear information of samples. Multiple kernel clustering provides an elegant framework for grouping samples into different categories by extracting complementary information from multiple sources. Through highly efficient and high-quality clustering, the efficiency of data analysis is greatly enhanced and labor costs are saved.
By fully considering the relationships between samples, a local kernel alignment variable has been developed. Experiments has proven that this can improve clustering performance. By assuming that an optimal kernel is located in the neighborhood of a combined kernel, an optimal neighborhood multiple kernel clustering algorithm has been provided that improves clustering performance by improving the representability of learned optimal kernels. The algorithm maximally aligns multiple base partitions with consistent partitions, with considerable algorithm acceleration and satisfactory clustering performance. On this basis, an effective algorithm based on late-fusion for processing incomplete multi-view data has been provided.
As a representative of multi-view clustering, a new simple multiple kernel k-means (SimpleMKKM) has recently been provided. SimpleMKKM does not jointly minimize the kernel weight and the clustering partition matrix, but minimizes the kernel weight and maximizes the clustering partition matrix optimization framework, resulting in a troublesome minimum-maximum optimization. It can then be equivalently converted into a minimization problem, and a reduced gradient algorithm is designed to solve the resulting optimization. This algorithm has proven to be an efficient optimization with robustness to noise perspectives and has attracted a wide range of attention from many researchers.
Although the recently provided SimleMKKM has the above-mentioned advantages, it has been observed that it tightly aligns the combined kernel matrix with the “ideal” similarity globally generated by the clustering partition matrix. This can force all sample pairs to align equally with the same ideal similarity indiscriminately. Therefore, it cannot effectively handle the relationships between samples, neglects local structures, and possibly results in unsatisfactory clustering performance.
The present application aims to address the defects of the prior art by providing a face image clustering method and system based on localized simple multiple kernel k-means.
In order to achieve the above objective, the present application adopts the following technical solutions:
Provided is a face image clustering method based on localized simple multiple kernel k-means, including the following steps:
Further, calculating the localized kernel matrix for each view in S3 is expressed as:
wherein
Further, the simple multiple kernel k-means clustering objective function in S4 is expressed as:
wherein γ represents a coefficient vector; H represents a partition matrix; HT represents a permutation of a commutation matrix; Kγ represents a combined kernel matrix of Kp generated by γ; Ik represents a k-order identity matrix.
Further, the localized simple multiple kernel k-means clustering objective function in S4 is expressed as:
wherein Δ={γ∈m|Σp=1mγp=1,γp≥0, ∀p}, Rm represents an m-dimensional real number vector space; γp represents the p-th component of γ.
Further, solving the minimum of the constructed objective function in S5 specifically includes: simplifying the localized simple multiple kernel k-means clustering objective function into a simple multiple kernel k-means clustering objective function:
wherein ⊕ represents the multiplication of elements; {tilde over (K)}p=(Σi=1nA(i))⊕Kp represents a standardized kernel matrix;
representing an optimum function.
Further, solving the minimum of the constructed objective function by adopting the reduced gradient descent method in S5 specifically includes:
wherein dp represents a descending direction.
Correspondingly, further provided is a face image clustering system based on localized simple multiple kernel k-means, including:
Further, calculating the localized kernel matrix for each view in the second calculation module is expressed as:
wherein
Further, the simple multiple kernel k-means clustering objective function in the construction module is expressed as:
wherein γ represents a coefficient vector; H represents a partition matrix; HT represents a permutation of a commutation matrix; Kγ represents a combined kernel matrix of Kp generated by γ; Ik represents a k-order identity matrix.
Further, the localized simple multiple kernel k-means clustering objective function in the construction module is expressed as:
Rm represents an m-dimensional real number vector space; γp represents the p-th component of γ.
Compared with the prior art, the present application provides a novel localized simple multiple kernel k-means clustering machine learning method, which includes modules of localized kernel alignment, objective function optimization to obtain an optimal combination coefficient γ, a corresponding partition matrix H, and the like. In the present application, through the optimization of the objective function, the optimized kernel combination can represent the information of a single view, and can better serve the view fusion, so that the purpose of improving the clustering effect is achieved. Moreover, the present application performs localization processing on each view to strengthen local information. MKKM-MM was the first attempt to improve MKKM through min-max learning, which indeed improves MKKM, but has a limited effect. The performance of the provided localized SimpleMKKM significantly surpasses that of MKKM-MM. This once again proves the advantages of the formula and the associated optimization strategy of the present application. The localized SimpleMKKM consistently and significantly outperforms SimpleMKKM.
The embodiments of the present application are illustrated below through specific examples, and other advantages and effects of the present application can be easily understood by those skilled in the art based on the contents disclosed herein. The present application can also be implemented or applied through other different specific embodiments. Various modifications or changes to the details in the description can be made based on different perspectives and applications without departing from the spirit of the present application. It should be noted that, unless conflicting, the embodiments and features of the embodiments below may be combined with each other.
The present application aims to address the defects of the prior art by providing a face image clustering method and system based on localized simple multiple kernel k-means.
The embodiment provides a face image clustering method based on localized simple multiple kernel k-means, as shown in
The kernel k-means clustering process is as follows: {xi}i=1n⊆=X is set as a dataset consisting of n samples, and ϕ(·):x|X→ represents a feature mapping that projects sample x into a reproducing kernel Hilbert space . The target of kernel k-means clustering is to minimize the sum of squared errors based on the partition matrix B∈{0,1}n×k, as shown in the following formula:
and nc=Σi=1nBic represents the number of samples belonging to the c-th cluster (1≤c≤k). The above formula can be reformulated as:
wherein K is a kernel matrix with elements Kij=ϕ(xi)Tϕ(xj), L=diag([n1−1, . . . , nk−1]), and 1s∈s represents a vector in which all elements are 1.
Since variable B in the above formula is discrete, optimization is difficult.
is set, and the discrete constraint is converted into a real-value orthogonal constraint, i.e., HTH=Ik. The target formula may be converted into:
the closed-form solution thereof is given by the eigenvectors corresponding to the k maximum eigenvalues of the kernel matrix K, which can be obtained by eigendecomposition of K.
In S3, the localized kernel matrix for each view is calculated according to the nearest neighbor matrices.
The localized kernel matrix for each view is expressed as:
wherein
In S4, the localized simple multiple kernel k-means clustering objective function is constructed according to the calculated localized kernel matrix for each view.
The simple multiple kernel k-means clustering objective function is expressed as:
wherein γ represents a coefficient vector; H represents a partition matrix; HT represents a permutation of a commutation matrix; Kγ represents a combined kernel matrix of Kp generated by γ; Ik represents a k-order identity matrix.
S(i)∈{0,1}n×round(τ×n) represents the (τ×n)-nearest neighbor indicator matrix of the i-th sample, and round(·) represents a rounding function. The embodiment defines the local alignment of the i-th sample, which is expressed as:
wherein S(i)
The localized simple multiple kernel k-means clustering objective function is expressed as:
Rm represents . . . ; γp represents . . . ; Kγ=Σp=1mγp2κp, A(i)=S(i)S(i)
In S5, the reduced gradient descent method is used to solve the minimum of the constructed objective function to obtain the optimal clustering partition matrix.
(1) The objective function of the simple multiple kernel k-means clustering SimpleMKKM is a special case of the localized simple multiple kernel k-means clustering objective function described above, then:
wherein ⊕ represents the multiplication of elements; {tilde over (K)}p=Σ(Σi=1nA(i))⊕Kp represents a standardized kernel matrix;
As can be seen from the above, the aforementioned formula is simplified to SimpleMKKM when all elements of A(i) are set to 1. In this case, each sample takes the remaining samples as its neighbors. This means that SimpleMKKM may be taken as a special case of the above formula, and therefore, the above formula can be written equally effectively as:
in this case, the min-max optimization is transformed into a min optimization, where its target (γ) is a kernel k-means optimum function.
(2) Through the above normalization, each {tilde over (K)}p still remains positive semi-definite (PSD).
This is specified with each k, being a positive semi-definite matrix.
Note that S(i)∈{0,1}n×round(τ×n) is a positive semi-definite matrix. Moreover, the product on the elements between two positive semi-definite matrices is still a positive semi-definite matrix. Therefore, each {tilde over (K)}p is a positive semi-definite matrix.
Every {tilde over (K)}p remains positive semi-definite through the above normalization, which guarantees the differentiability of (γ). The differentiability of (γ) is demonstrated next below, showing how its gradient is calculated and optimized using a reduced gradient descent algorithm.
(3) (γ) is differentiable, and the global uniqueness of (γ) can be derived.
The objective function is calculated by the gradient descent method as:
u is set as a number indicating the largest component of vector γ, which is believed to provide better numerical stability.
The embodiment considers a positive constraint of γ in a descending direction, which is expressed as:
wherein dp represents the descending direction, and γ can be calculated by γ←γ+αd, where α is the optimal step length. It may be selected by a one-dimensional line search strategy, such as an Armijo criterion.
The embodiment discusses the computational complexity of the provided localized simple MKKM. In each iteration, the localized SimpleMKKM needs to solve a kernel k-means problem, calculate the descending direction, and search for the optimal step length. Therefore, its computational complexity at (n3+m*n3+m*n0), where no is the maximum operand required to find the optimal step length. As observed, the localized simple MKKM does not significantly increase the computational complexity of existing MKKM and SimpleMKKM algorithms.
The convergence of the localized SimpleMKKM is then briefly discussed. Note that with a given γ, it becomes a traditional kernel k-means, which has a global optimum. Under this condition, the gradient calculation in step (3) is precise, the algorithm of the embodiment performs reduced gradient descent on a domain, and the function converges to the minimum of (γ).
The embodiment provides a novel localized simple multiple kernel k-means clustering machine learning method, which includes modules of localized kernel alignment, objective function optimization to obtain an optimal combination coefficient γ, a corresponding partition matrix H, and the like. In the embodiment, through the optimization of the objective function, the optimized kernel combination can represent the information of a single view, and can better serve the view fusion, so that the purpose of improving the clustering effect is achieved. Moreover, the embodiment performs localization processing on each view to strengthen local information. MKKM-MM was the first attempt to improve MKKM through min-max learning, which indeed improves MKKM, but has a limited effect. The performance of the provided localized SimpleMKKM significantly surpasses that of MKKM-MM. This once again proves the advantages of the formula and the associated optimization strategy of the embodiment. The localized SimpleMKKM consistently and significantly outperforms SimpleMKKM.
Correspondingly, further provided is a face image clustering system based on localized simple multiple kernel k-means, including:
Further, calculating the localized kernel matrix for each view in the second calculation module is expressed as:
wherein
Further, the simple multiple kernel k-means clustering objective function in the construction module is expressed as:
wherein γ represents a coefficient vector; H represents a partition matrix; HT represents a permutation of a commutation matrix; Kγ represents a combined kernel matrix of Kp generated by γ; Ik represents a k-order identity matrix.
Further, the localized simple multiple kernel k-means clustering objective function in the construction module is expressed as:
Rm represents an m-dimensional real number vector space; γp represents the p-th component of γ.
The difference between the face image clustering method based on localized simple multiple kernel k-means provided in the embodiment and that provided in Embodiment I is as follows: The embodiment tested the clustering performance of the present method on 8 MKKM benchmark datasets, including MSRA, Still, Cal-7, PFold, Nonpl, Flo17, Flo102, and Reuters. See table 1 for relevant information on the datasets.
The embodiment adopted an average multiple kernel k-means (A-MKKM) clustering algorithm, multiple kernel k-means (MKKM) clustering, localized multiple kernel k-means (LMKKM) clustering, robust multiple kernel k-means clustering (MKKM-MM), multiple kernel k-means clustering with matrix-induced regularization (MKKM-MR), optimal neighborhood multiple kernel clustering (ONKC), multi-view clustering via late fusion alignment maximization (MVC-LFA), and multiple kernel clustering with local alignment maximization (LKAM). In all experiments, all benchmark kernels were first centered and regularized. For all datasets, the number of categories was assumed to be known and set as the number of clustering categories. In addition, in the embodiment, grid search was used for the parameters of RMKKM, MKKM-MR, ONKC, and MVC-LFA.
The embodiment used common cluster accuracy (ACC), normalized mutual information (NMI), and Rand index (RI) to show the clustering performance of each method. All methods were randomly initialized and repeated 50 times and showed the best results to reduce the randomness caused by k-means.
indicates data missing or illegible when filed
Table 2 shows the clustering effects of the above methods and the comparison algorithms on all datasets. From this table, it can be observed that: 1. MKKM-MM was the first attempt to improve MKKM through min-max learning. As observed, it does improve MKKM, but the performance improvement over MKKM is limited across all datasets. Meanwhile, the performance of the provided localized simple MKKM significantly surpasses that of MKKM-MM. This once again proves the advantages of the method and the associated optimization strategy of the embodiment. 2. In addition to the localized SimpleMKKM of the method, SimpleMKKM also achieves comparable or better clustering performance than the above algorithms on all benchmark datasets. The superiority is attributed to its new formula and new optimization algorithm. 3. The provided localized SimpleMKKM consistently and significantly outperforms SimpleMKKM. For example, ACC exceeds SimpleMKKM algorithm by 4.7%, 5.2%, 8.3%, 1.2%, 17.3%, 17.3%, 1.8%, 1.5%, and 1.1% on 8 benchmark datasets. Improvements in other criteria are also similar. These results well prove the superiority of the provided localized simple MKKM benefiting from exploring and extracting local information of the kernel matrix.
The embodiment provides a novel localized simple multiple kernel k-means clustering machine learning method, which includes modules of localized kernel alignment, objective function optimization to obtain an optimal combination coefficient γ, a corresponding partition matrix H, and the like. In the embodiment, through the optimization of the objective function, the optimized kernel combination can represent the information of a single view, and can better serve the view fusion, so that the purpose of improving the clustering effect is achieved. Moreover, the embodiment performs localization processing on each view to strengthen local information. MKKM-MM was the first attempt to improve MKKM through min-max learning, which indeed improves MKKM, but has a limited effect. The performance of the provided localized SimpleMKKM significantly surpasses that of MKKM-MM. This once again proves the advantages of the formula and the associated optimization strategy of the embodiment. The localized SimpleMKKM consistently and significantly outperforms SimpleMKKM.
It should be noted that the above descriptions are only preferred embodiments of the present application and the principles of the employed technologies. It should be understood by those skilled in the art that the present application is not limited to the particular embodiments described herein, and those skilled in the art can make various obvious changes, rearrangements, and substitutions without departing from the protection scope of the present application. Therefore, although the present application has been described in more detail by the above embodiments, it is not limited to the above embodiments, and may further include other equivalent embodiments without departing from the concept of the present application, and the scope of the present application is determined by the scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
202110940777.6 | Aug 2021 | CN | national |
This application is the national phase entry of International Application No. PCT/CN2022/112016, filed on Aug. 12, 2022, which is based upon and claims priority to Chinese Patent Application No. 202110940777.6, filed on Aug. 17, 2021, the entire contents of which are incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CN2022/112016 | 8/12/2022 | WO |