GATES: Using Graph Attention Networks for Entity Summarization

The sheer size of modern knowledge graphs has led to increased attention being paid to the entity summarization task. Given a knowledge graph T and an entity e found therein, solutions to entity summarization select a subset of the triples from T which summarize e's concise bound description. Presently, the best performing approaches rely on sequence-to-sequence models to generate entity summaries and use little to none of the structure information of T during the summarization process. We hypothesize that this structure information can be exploited to compute better summaries. To verify our hypothesis, we propose GATES, a new entity summarization approach that combines topological information and knowledge graph embeddings to encode triples. The topological information is encoded by means of a Graph Attention Network. Furthermore, ensemble learning is applied to boost the performance of triple scoring. We evaluate GATES on the DBpedia and LMDB datasets from ESBM (version 1.2), as well as on the FACES datasets. Our results show that GATES outperforms the state-of-the-art approaches on 4 of 6 configuration settings and reaches up to 0.574 F-measure. Pertaining to resulted summaries quality, GATES still underperforms the state of the arts as it obtains the highest score only on 1 of 6 configuration settings at 0.697 NDCG score. An open-source implementation of our approach and of the code necessary to rerun our experiments are available at https://github.com/dice-group/GATES.

[1]  W. Bruce Croft,et al.  A Deep Relevance Matching Model for Ad-hoc Retrieval , 2016, CIKM.

[2]  Yuzhong Qu,et al.  RELIN: Relatedness and Informativeness-Based Centrality for Entity Summarization , 2011, International Semantic Web Conference.

[3]  Axel-Cyrille Ngonga Ngomo,et al.  GENESIS: a generic RDF data access interface , 2017, WI.

[4]  Jason Weston,et al.  Translating Embeddings for Modeling Multi-relational Data , 2013, NIPS.

[5]  Gong Cheng,et al.  Entity Summarization with User Feedback , 2020, ESWC.

[6]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[7]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[8]  Amit P. Sheth,et al.  Gleaning Types for Literals in RDF Triples with Application to Entity Summarization , 2016, ESWC.

[9]  Claudia Beleites,et al.  Assessing and improving the stability of chemometric models in small sample size situations , 2008, Analytical and bioanalytical chemistry.

[10]  Yaxin Liu,et al.  ESA: Entity Summarization with Attention , 2019, ArXiv.

[11]  Austin Melton,et al.  Automatic Weight Generation and Class Predicate Stability in RDF Summary Graphs , 2015, IESD@ISWC.

[12]  Wei Zhou,et al.  AutoSUM: Automating Feature Extraction and Multi-user Preference Simulation for Entity Summarization , 2020, PAKDD.

[13]  Weiwei Xing,et al.  STGAT: Spatial-Temporal Graph Attention Networks for Traffic Flow Forecasting , 2020, IEEE Access.

[14]  R. Glynn,et al.  The Wilcoxon Signed Rank Test for Paired Comparisons of Clustered Data , 2006, Biometrics.

[15]  Matthijs Douze,et al.  FastText.zip: Compressing text classification models , 2016, ArXiv.

[16]  Alex Graves,et al.  Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.

[17]  Marcin Sydow,et al.  DIVERSUM: Towards diversified summarisation of entities in knowledge graphs , 2010, 2010 IEEE 26th International Conference on Data Engineering Workshops (ICDEW 2010).

[18]  Yoelle Maarek,et al.  An incremental conceptual clustering algorithm that reduces input-ordering bias , 1990 .

[19]  Nelia Lasierra,et al.  LinkSUM: Using Link Analysis to Summarize Entity Data , 2016, ICWE.

[20]  Zhen Wang,et al.  Knowledge Graph Embedding by Translating on Hyperplanes , 2014, AAAI.

[21]  Rui Zhang,et al.  Graph-based Neural Multi-Document Summarization , 2017, CoNLL.

[22]  Haopeng Chen,et al.  EGAT: Edge-Featured Graph Attention Network , 2021, ICANN.

[23]  Axel-Cyrille Ngonga Ngomo,et al.  Convolutional Complex Knowledge Graph Embeddings , 2021, ESWC.

[24]  Yoshua Bengio,et al.  A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..

[25]  Guillaume Bouchard,et al.  Complex Embeddings for Simple Link Prediction , 2016, ICML.

[26]  Rajeev Motwani,et al.  The PageRank Citation Ranking : Bringing Order to the Web , 1999, WWW 1999.

[27]  Gong Cheng,et al.  ESBM: An Entity Summarization BenchMark , 2020, ESWC.

[28]  Jianfeng Gao,et al.  Embedding Entities and Relations for Learning and Inference in Knowledge Bases , 2014, ICLR.

[29]  Bernhard Schölkopf,et al.  Learning with Hypergraphs: Clustering, Classification, and Embedding , 2006, NIPS.

[30]  Gong Cheng,et al.  DeepLENS: Deep Learning for Entity Summarization , 2020, DL4KG@ESWC.

[31]  Yuzhong Qu,et al.  Entity Summarization: State of the Art and Future Challenges , 2021, J. Web Semant..

[32]  Amit P. Sheth,et al.  FACES: Diversity-Aware Entity Summarization Using Incremental Hierarchical Conceptual Clustering , 2015, AAAI.