Abstract
This chapter introduces three hypergraph computation paradigms, including intra-hypergraph computation, inter-hypergraph computation, and hypergraph structure computation. Intra-hypergraph computation representation aims to conduct representation learning of a hypergraph, where each subject is represented by a hypergraph of its components. Inter-hypergraph computation is to conduct representation learning of vertices in the hypergraph, where each subject is a vertex in the hypergraph. Hypergraph structure computation is to conduct hypergraph structure prediction, which aims to find the connections among vertices. This chapter is a general introduction of hypergraph computation paradigms to show how to formulate the task in the hypergraph computation framework.
You have full access to this open access chapter, Download chapter PDF
3.1 Introduction
Hypergraph computation can be roughly divided into three types: representation learning of a hypergraph, where each subject is represented by a hypergraph of its components, representation learning of vertices in the hypergraph, where each subject is a vertex in the hypergraph, and hypergraph structure prediction, which aims to find the connections among vertices. There are three types of computation paradigms that can be named intra-hypergraph computation, inter-hypergraph computation, and hypergraph structure computation. In this chapter, we introduce the generalized computation paradigms corresponding to these three directions and show how to formulate practical tasks in these hypergraph computation frameworks. We note that specific implementations of generalized functions in the paradigm are not introduced here, as they are parts of specifically defined functions or modules in the hypergraph computation framework and will be introduced in subsequent chapters.
3.2 Intra-hypergraph Computation
Intra-hypergraph computation targets on learning the representation of a single subject using the inside component information, in which the correlations among the components of this subject are formulated in a hypergraph. In this hypergraph, the components of this subject are regarded by the set of vertices, and their high-order correlations are modeled by hyperedges. In this way, the individual subject is transformed into a hypergraph. As this hypergraph is generated by the subject’s components themselves, we can name this hypergraph as the intra-hypergraph of this subject.
Image representation and understanding [1,2,3] are typical intra-hypergraph computation applications. For example, an image can be split into a group of patches, and each patch is denoted by a vertex in the hypergraph. The hypergraph can be generated according to the semantic and spatial information of these patches. The information of these patches and their high-order correlations can be then used simultaneously to learn the representation for the image.
The general paradigm of intra-hypergraph computation can be described as follows. Given a target subject that contains n components, that are represented by feature vectors \(\mathbf {X}\in \mathbb {R}^{n \times d}\). An intra-hypergraph \(\mathbb {G}\) can be generated to formulate the high-order correlations inside the subject, whose incidence matrix is denoted by H. The representation of the individual subject can be learned by
where Θ denotes the to-be-learned parameters. The function f Θ(⋅) can be the neural network layers or other computing operators that aggregate the information of vertices together based on the hypergraph structure. Intra-hypergraph computation integrates the complex correlations among components into the learned representation, which can extract more information than simple aggregation operations.
In this paradigm, the subject to be analyzed is regarded as a whole system, and the intra-hypergraph is to model the correlation inside the system. This process is shown in Fig. 3.1.
3.3 Inter-hypergraph Computation
Inter-hypergraph computation targets at learning the representation of a subject by considering its correlations with other subjects. In this hypergraph, each subject, including the target one, is regarded by the set of vertices, and their high-order correlations are modeled by hyperedges. In this way, this group of subjects is transformed into a hypergraph. As this hypergraph is generated by the cross-subject correlations, we can name this hypergraph as the inter-hypergraph of this subject. Subject classification and retrieval [4,5,6,7] are typical inter-hypergraph computation applications. For example, we take an image as the target subject, and we can also have a pool of images for processing. Each image can be denoted by a vertex in the hypergraph. The hypergraph can be generated according to the semantic and spatial information of these images. The information of these images and their high-order correlations can be then used simultaneously to learn the representation of the target image.
The general paradigm of inter-hypergraph computation can be described as follows. Given a target subject and other n − 1 subjects, represented by feature vectors \(\mathbf {X}\in \mathbb {R}^{n \times d}\), an inter-hypergraph \(\mathbb {G}\) can be generated to formulate the high-order correlations among these subjects, whose incidence matrix is denoted by H. The representation of the target subject can be learned by
The vertex embedding can be further used for the downstream tasks, such as vertex classification, where the vertices are associated with pre-defined labels Y ∈ [K]n. This process is also shown in Fig. 3.1.
It is noted that a hypergraph structure can be either homogeneous or heterogeneous, depending on the definition of vertices. Given multiple types of data, or multi-modal data, another way to formulate such correlations is to generate multiple hypergraphs accordingly. For example, supposing that there are m types of features or modalities, denoted by X 1, X 2, …, X m, we can construct one hypergraph for each modality respectively. In this way, we can have m hypergraphs \(\mathbb {G}_1 = (\mathbb {V}_1; \mathbb {E}_1; {\mathbf {W}}_1); \mathbb {G}_2 = (\mathbb {V}_2; \mathbb {E}_2; {\mathbf {W}}_2); ... ;\mathbb {G}_m = (\mathbb {V}_m; \mathbb {E}_m; {\mathbf {W}}_m)\) for the data with m modalities. The general paradigm for multi-modal inter-hypergraph computation can be described as
where H 1, H 2, …, H m are the incidence matrices of the m hypergraphs.
3.4 Hypergraph Structure Computation
Hypergraph structure computation aims to learn the high-order correlations among data in the presence of missing links and inaccurate initial structure. There are two scenarios in which hypergraph structure computation is performed: either the set of hyperedges is incomplete or the affiliation relationships between vertices and hyperedges are incomplete. Recommender system and drug discovery [8,9,10] are typical hypergraph structure computation applications. For example, in recommender system, the hyperedges describe the connections between items and users with specific semantics. The number of hyperedges is fixed, and the features of both vertices and hyperedges can be obtained as the input. Here, the target of hypergraph structure computation is to predict whether a vertex belongs to a hyperedge or not. If a new hyperedge is predicted, we can have new link to indicate the connections. However, in a knowledge hypergraph, the hyperedges display the facts in the real world, which are usually highly incomplete. The missing links are expected to be inferred based on existing links by hypergraph structure computation. Therefore, in the second case, the objective of hypergraph structure computation is not only optimizing existing links but also inferring the unobserved links.
In the following, we describe the computation paradigms of these two cases separately. The first scenario is that the set of hyperedges is complete and the affiliation relationships between vertices and hyperedges are incomplete. In this case, we usually can extract a feature vector for each hyperedge representation. Given the input of vertex features \({\mathbf {X}}_{\mathbb {V}}\) and hyperedge features \({\mathbf {X}}_{\mathbb {E}}\), we can calculate the incidence matrix by the function related to the vertex and hyperedge features as
For example, the attention score can be used as an instance of the function in practice.
In the second scenario, if there are missing hyperedges in the observed hypergraph and the semantics of hyperedges are ambiguous, it is difficult to directly describe the hyperedges by features. Consequently, only the initial incomplete hypergraph structure and the features of vertices can be available as the input. We denote the incidence matrix of the initial hypergraph structure by H (0). The computation paradigm can be written as
which indicates that the new hypergraph structure is updated based on the original hypergraph structure following specific prior information.
To guide the evolution of hypergraph structure to more accurately model data correlation, it is necessary to evaluate the quality of hypergraph structure based on the training data and prior information. If there is part of ground truth information about the hypergraph structure, the performance of correlation modeling can be evaluated directly. However, there is no golden standard for hypergraph structure in most cases. Therefore, we may need to perform downstream tasks using the new hypergraph and indirectly evaluate hypergraph computation performance through the downstream task results. Here, we refer to Fig. 3.1, and hypergraph structure computation can be conducted under the intra- and inter-hypergraph computation frameworks.
3.5 Summary
In this chapter, we introduce three hypergraph computation paradigms for different scenarios. These three paradigms are intra-hypergraph computation, inter-hypergraph computation, and hypergraph structure computation, which focus on learning the representation of a single subject using the inside component information, learning the representation of a subject by considering its correlations with other subjects, and learning the high-order correlations among data in the presence of missing links and inaccurate initial structure. This chapter provides an overview of how to use hypergraph computation, and the detailed hypergraph computation theory, methods, and application will be introduced in the following chapters.
References
D. Di, S. Li, J. Zhang, Y. Gao, Ranking-Based survival prediction on histopathological whole-slide images, in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (2020), pp. 428–438
D. Di, J. Zhang, F. Lei, Q. Tian, Y. Gao, Big-hypergraph factorization neural network for survival prediction from whole slide image. IEEE Trans. Image Process. 31, 1149–1160 (2022)
D. Di, C. Zou, Y. Feng, H. Zhou, R. Ji, Q. Dai, Y. Gao, Generating hypergraph-based high-order representations of whole-slide histopathological images for survival prediction. IEEE Trans. Pattern Analy. Mach. Intell. 1–16 (2022). https://doi.org/10.1109/TPAMI.2022.3209652
Y. Gao, Z. Zhang, H. Lin, X. Zhao, S. Du, C. Zou, Hypergraph learning: methods and practices. IEEE Trans. Pattern Analy. Mach. Intell. 44(5), 2548–2566 (2022)
Y. Gao, M. Wang, D. Tao, R. Ji, Q. Dai, 3-D object retrieval and recognition with hypergraph analysis. IEEE Trans. Image Process. 21, 4290–4303 (2012)
Y. Feng, H. You, Z. Zhang, R. Ji, Y. Gao, Hypergraph neural networks, in Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (2019), pp. 3558–3565
Y. Gao, M. Wang, Z.J. Zha, J. Shen, X. Li, X. Wu, Visual textual joint relevance learning for tag-based social image search. IEEE Trans. Image Process. 22(1), 363–376 (2013)
S. Ji, Y. Feng, R. Ji, X. Zhao, W. Tang, Y. Gao, Dual channel hypergraph collaborative filtering, in Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (2020), pp. 2020–2029
H. Fan, F. Zhang, Y. Wei, Z. Li, C. Zou, Y. Gao, Q. Dai, Heterogeneous hypergraph variational autoencoder for link prediction. IEEE Trans. Pattern Analy. Mach. Intell. 44(8), 4125–4138 (2021)
D. Ruan, S. Ji, C. Yan, J. Zhu, X. Zhao, Y. Yang, Y. Gao, C. Zou, Q. Dai, Exploring complex and heterogeneous correlations on hypergraph for the prediction of drug-target interactions. Patterns 2(12), 100390 (2021)
Author information
Authors and Affiliations
Rights and permissions
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
Copyright information
© 2023 The Author(s)
About this chapter
Cite this chapter
Dai, Q., Gao, Y. (2023). Hypergraph Computation Paradigms. In: Hypergraph Computation. Artificial Intelligence: Foundations, Theory, and Algorithms. Springer, Singapore. https://doi.org/10.1007/978-981-99-0185-2_3
Download citation
DOI: https://doi.org/10.1007/978-981-99-0185-2_3
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0184-5
Online ISBN: 978-981-99-0185-2
eBook Packages: Computer ScienceComputer Science (R0)