| Literature DB >> 35615554 |
Zhe Yang1,2,3, Liangkui Xu1,2,3, Lei Zhao1,2,3.
Abstract
Hypergraph learning is a new research hotspot in the machine learning field. The performance of the hypergraph learning model depends on the quality of the hypergraph structure built by different feature extraction methods as well as its incidence matrix. However, the existing models are all hypergraph structures built based on one feature extraction method, with limited feature extraction and abstract expression ability. This paper proposed a multimodal feature fusion method, which firstly built a single modal hypergraph structure based on different feature extraction methods, and then extended the hypergraph incidence matrix and weight matrix of different modals. The extended matrices fuse the multimodal abstract feature and an expanded Markov random walk range during model learning, with stronger feature expression ability. However, the extended multimodal incidence matrix has a high scale and high computational cost. Therefore, the Laplacian matrix fusion method was proposed, which performed Laplacian matrix transformation on the incidence matrix and weight matrix of every model, respectively, and then conducted a weighted superposition on these Laplacian matrices for subsequent model training. The tests on four different types of datasets indicate that the hypergraph learning model obtained after multimodal feature fusion has a better classification performance than the single modal model. After Laplace matrix fusion, the average time can be reduced by about 40% compared with the extended incidence matrix, the classification performance can be further improved, and the index F1 can be improved by 8.4%.Entities:
Mesh:
Year: 2022 PMID: 35615554 PMCID: PMC9126688 DOI: 10.1155/2022/9073652
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Hypergraph and incidence matrix. (a) Hypergraph. (b) Incidence matrix.
Figure 2Hypergraph matrix extension (denotes blend operation).
Classification performance of modal combinations.
| Datasets | Modal combinations | Incidence matrix extension | Incidence matrix extension & Laplacian matrix fusion | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Accuracy | Precision | Recall | F1 | Accuracy | Precision | Recall | F1 | ||
| Cat & Dog | Poor + Poor | 0.687 | 0.662 | 0.674 | 0.667 | 0.722 | 0.718 | 0.720 | 0.718 |
| PHA + SIFT | |||||||||
| Poor + Good | 0.919 | 0.920 | 0.919 | 0.919 | 0.928 | 0.926 | 0.930 | 0.928 | |
| PHA + RVGG | |||||||||
| Good + Good | 0.975 | 0.975 | 0.977 | 0.976 | 0.986 | 0.986 | 0.986 | 0.986 | |
| RVGG + HVGG | |||||||||
|
| |||||||||
| Ctrip | Poor + Poor | 0.619 | 0.620 | 0.619 | 0.619 | 0.623 | 0.621 | 0.623 | 0.621 |
| Jaccard + TF-IDF | |||||||||
| Good + Poor | 0.646 | 0.821 | 0.613 | 0.701 | 0.648 | 0.837 | 0.624 | 0.711 | |
| TF-IDF + Doc2vec | |||||||||
| Good + Good | 0.659 | 0.878 | 0.709 | 0.785 | 0.722 | 0.718 | 0.720 | 0.718 | |
| Doc2vec + word2vec | |||||||||
Figure 3Laplacian matrix fusion.
Test datasets.
| Datasets | Datasets type | Total samples | Numbers of samples in training set | Categories |
|---|---|---|---|---|
| Cat & Dog | Image | 2000 | 200 | 2 |
| Cifar 10 | Image | 40000 | 6000 | 10 |
| Ctrip | Text | 7766 | 700 | 2 |
| Spambase | Numerical value | 4601 | 400 | 2 |
Classification performance of single modal hypergraph model.
| Datasets | Feature extraction method | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|---|
| Cat & Dog | PHA | 0.552 | 0.554 | 0.574 | 0.563 |
| SIFT | 0.658 | 0.660 | 0.663 | 0.661 | |
| HSIFT | 0.669 | 0.674 | 0.678 | 0.675 | |
| VGG | 0.958 | 0.960 | 0.957 | 0.958 | |
| ResNet | 0.960 | 0.961 | 0.958 | 0.959 | |
| HVGG | 0.963 | 0.962 | 0.963 | 0.962 | |
| RVGG | 0.965 | 0.965 | 0.965 | 0.965 | |
|
| |||||
| Ctrip | Jaccard | 0.508 | 0.505 | 0.513 | 0.508 |
| TF-IDF | 0.603 | 0.609 | 0.613 | 0.610 | |
| LSI | 0.608 | 0.612 | 0.620 | 0.615 | |
| Word2vec | 0.639 | 0.748 | 0.651 | 0.696 | |
| Doc2vec | 0.651 | 0.870 | 0.657 | 0.748 | |
Figure 4Influence of modal quantity on model performance. (a) Cat & Dog dataset. (b) Ctrip dataset.
Time cost comparison.
| Datasets | Model | Accuracy | Precision | Recall | F1 | Time (s) |
|---|---|---|---|---|---|---|
| Cat & Dog | RVGG | 0.965 | 0.965 | 0.965 | 0.965 | 20.26 |
| RVGG + HVGG | 0.975 | 0.975 | 0.977 | 0.976 | 49.04 | |
| Incidence matrix extension | ||||||
| RVGG + HVGG | 0.986 | 0.986 | 0.986 | 0.986 | 27.35 | |
| Incidence matrix extension + Laplacian matrix fusion | ||||||
|
| ||||||
| Cifar 10 | RVGG | 0.561 | 0.570 | 0.564 | 0.567 | 11236 |
| RVGG + HVGG | 0.564 | 0.573 | 0.573 | 0.573 | 23793 | |
| Incidence matrix extension | ||||||
| RVGG + HVGG | 0.594 | 0.613 | 0.583 | 0.598 | 13158 | |
| Incidence matrix extension + Laplacian matrix fusion | ||||||
|
| ||||||
| Ctrip | Doc2vec | 0.651 | 0.870 | 0.657 | 0.748 | 59.7 |
| Doc2vec + word2vec | 0.659 | 0.878 | 0.709 | 0.785 | 106.86 | |
| Incidence matrix extension | ||||||
| Doc2vec + word2vec | 0.663 | 0.884 | 0.721 | 0.851 | 74.2 | |
| Incidence matrix extension + Laplacian matrix fusion | ||||||
|
| ||||||
| Spambase | Euclidean | 0.646 | 0.701 | 0.646 | 0.672 | 8.17 |
| Cosin + Euclidean incidence matrix extension | 0.654 | 0.729 | 0.654 | 0.689 | 17.3 | |
| Cosin + Euclidean incidence matrix extension + Laplacian matrix fusion | 0.712 | 0.734 | 0.696 | 0.714 | 10.25 | |
Figure 5Comparison of model performance. (a) Cat & Dog dataset. (b) Ctrip dataset.