The paper describes an innovative approach to expanding the domain coverage of the Slovene wordnet (sloWNet) by exploiting multiple resources. In the experiment described here we are using a large monolingual Slovene corpus of texts from the domain of informatics to harvest terminology from, and a parallel English-Slovene corpus and an online dictionary as bilingual resources to facilitate the mapping of terms to sloWNet. We first identify the core terms of the domain in English using the Princeton University's WordNet 2.1, and then we translate them into Slovene using a bilingual lexicon produced from the parallel corpus. In the next step we extract multi-word terms from the Slovene domain-specific corpus using a hybrid approach, and final...
Various efforts have been made for the development of tools and methods dedicated to the automatic p...
International audienceOne of the main tasks of the Natural Language Processing Group at the Faculty ...
The availability of large collections of text (language corpora) is crucial for empirically supporte...
sloWNet is the Slovene WordNet developed in the expand approach: it contains the complete Princeton ...
International audienceIn this paper we present an automatic, language-independent approach to extend...
International audienceIn this paper we present a language-independent, fully modular and automatic a...
International audienceThis paper compares automatically generated sets of synonyms in French and Slo...
Open Slovene WordNet (OSWN) is derived from Open English WordNet (https://en-word.net/), which itsel...
We present the current implementation state of our work consisting in interlinking language data and...
This is an automatically created Slovene thesaurus from Slovene data available in a comprehensive En...
In this paper we present a set of tools that will help developers of wordnets not only to in-crease ...
The paper describes the project whose main purpose is the creation of the Slovene terminology web po...
In this paper we present the mapping between WordNet domains and WordNet topics, and the emergent Wi...
In many scientific, technological or political fields terminology and the production of up-to- date ...
The objective of this paper is to present a method to automatically enrich WordNet with sub-trees of...
Various efforts have been made for the development of tools and methods dedicated to the automatic p...
International audienceOne of the main tasks of the Natural Language Processing Group at the Faculty ...
The availability of large collections of text (language corpora) is crucial for empirically supporte...
sloWNet is the Slovene WordNet developed in the expand approach: it contains the complete Princeton ...
International audienceIn this paper we present an automatic, language-independent approach to extend...
International audienceIn this paper we present a language-independent, fully modular and automatic a...
International audienceThis paper compares automatically generated sets of synonyms in French and Slo...
Open Slovene WordNet (OSWN) is derived from Open English WordNet (https://en-word.net/), which itsel...
We present the current implementation state of our work consisting in interlinking language data and...
This is an automatically created Slovene thesaurus from Slovene data available in a comprehensive En...
In this paper we present a set of tools that will help developers of wordnets not only to in-crease ...
The paper describes the project whose main purpose is the creation of the Slovene terminology web po...
In this paper we present the mapping between WordNet domains and WordNet topics, and the emergent Wi...
In many scientific, technological or political fields terminology and the production of up-to- date ...
The objective of this paper is to present a method to automatically enrich WordNet with sub-trees of...
Various efforts have been made for the development of tools and methods dedicated to the automatic p...
International audienceOne of the main tasks of the Natural Language Processing Group at the Faculty ...
The availability of large collections of text (language corpora) is crucial for empirically supporte...