In this paper, we present work on enhancing the basic data resource of a context-aware system. First, we introduce a supervised approach to extracting geographical relations on a fine-grained level. Second, we present a novel way of using Wikipedia as a corpus based on self-annotation. A self-annotation is an automatically created high-quality annotation that can be used for training and evaluation. The fined-grained relations are used to complete gazetteer data. The precision and recall scores of more than 97 % confirm that a statistical IE pipeline can be used to improve the data quality of community-based resources. 1
Abstract. Textual patterns have been used effectively to extract information from large text collect...
Wikification, commonly referred to as Disam-biguation to Wikipedia (D2W), is the task of identifying...
Social media sources such as Flickr and Twitter continuously generate large amounts of textual infor...
This paper addresses the challenge of extracting geospa-tial data from the article text of the Engli...
Abstract. The exponential growth of Wikipedia recently attracts the attention of a large number of r...
We provide a subcorpus of Wikipedia that was annotated with Wikidata relations using a distant super...
Ground-truth datasets are essential for the training and evaluation of any automated algorithm. As s...
The exponential growth and reliability of Wikipedia have made it a promising data source for intelli...
We present a minimally-supervised approach for learning part whole relations from texts. Unlike prev...
Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extrac...
Abstract. This paper reports University of Pittsburgh’s participation in GeoCLEF 2008. As the first ...
Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extrac...
We present a simple but effective method of automatically extracting domain-specific terms using Wik...
Proceedings of the First Workshop on Semantic Wikis - From Wiki to Semantics co-located with the ESW...
This thesis focuses on the design of algorithms for the extraction of knowledge (in terms of entitie...
Abstract. Textual patterns have been used effectively to extract information from large text collect...
Wikification, commonly referred to as Disam-biguation to Wikipedia (D2W), is the task of identifying...
Social media sources such as Flickr and Twitter continuously generate large amounts of textual infor...
This paper addresses the challenge of extracting geospa-tial data from the article text of the Engli...
Abstract. The exponential growth of Wikipedia recently attracts the attention of a large number of r...
We provide a subcorpus of Wikipedia that was annotated with Wikidata relations using a distant super...
Ground-truth datasets are essential for the training and evaluation of any automated algorithm. As s...
The exponential growth and reliability of Wikipedia have made it a promising data source for intelli...
We present a minimally-supervised approach for learning part whole relations from texts. Unlike prev...
Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extrac...
Abstract. This paper reports University of Pittsburgh’s participation in GeoCLEF 2008. As the first ...
Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extrac...
We present a simple but effective method of automatically extracting domain-specific terms using Wik...
Proceedings of the First Workshop on Semantic Wikis - From Wiki to Semantics co-located with the ESW...
This thesis focuses on the design of algorithms for the extraction of knowledge (in terms of entitie...
Abstract. Textual patterns have been used effectively to extract information from large text collect...
Wikification, commonly referred to as Disam-biguation to Wikipedia (D2W), is the task of identifying...
Social media sources such as Flickr and Twitter continuously generate large amounts of textual infor...