Semantic representation of unstructured text is crucial in modern artificial intelligence and information retrieval applications. The semantic information extraction process from an unstructured text fragment to a corresponding representation from a concept ontology is known as named entity disambiguation. In this work, we introduce a distributed, supervised deep learning methodology employing a long short-term memory-based deep learning architecture model for entity linking with Wikipedia. In the context of a frequently changing online world, we introduce and study the domain of online training named entity disambiguation, featuring on-the-fly adaptation to underlying knowledge changes. Our novel methodology evaluates polysemous anchor mentions with sense compatibility based on thematic segmentation of the Wikipedia knowledge graph representation. We aim at both robust performance and high entity-linking accuracy results. The introduced modeling process efficiently addresses conceptualization, formalization, and computational challenges for the online training entity-linking task. The novel online training concept can be exploited for wider adoption, as it is considerably beneficial for targeted topic, online global context consensus for entity disambiguation.
[1]
Roberto Cornacchia,et al.
Ranking Triples using Entity Links in a Large Web Crawl - The Chicory Triple Scorer at WSDM Cup 2017
,
2017,
ArXiv.
[2]
Christos Makris,et al.
Text Semantic Annotation: A Distributed Methodology Based on Community Coherence
,
2020,
Algorithms.
[3]
David Yarowsky,et al.
A method for disambiguating word senses in a large corpus
,
1992,
Comput. Humanit..
[4]
Roberto Navigli,et al.
Word sense disambiguation: A survey
,
2009,
CSUR.
[5]
Hamid Hassanpour,et al.
A comparative performance analysis of different activation functions in LSTM networks for classification
,
2017,
Neural Computing and Applications.
[6]
Jürgen Schmidhuber,et al.
Long Short-Term Memory
,
1997,
Neural Computation.