Back to Nlp Progress

Taxonomy Learning

english/taxonomy_learning.md

0.312.8 KB
Original Source

Taxonomy Learning

Taxonomy learning is the task of hierarchically classifying concepts in an automatic manner from text corpora. The process of building taxonomies is usually divided into two main steps: (1) extracting hypernyms for concepts, which may constitute a field of research in itself (see Hypernym Discovery below) and (2) refining the structure into a taxonomy.

Hypernym Discovery

Given a corpus and a target term (hyponym), the task of hypernym discovery consists of extracting a set of its most appropriate hypernyms from the corpus. For example, for the input word “dog”, some valid hypernyms would be “canine”, “mammal” or “animal”.

SemEval 2018

The SemEval-2018 hypernym discovery evaluation benchmark (Camacho-Collados et al. 2018), which can be freely downloaded here, contains three domains (general, medical and music) and is also available in Italian and Spanish (not in this repository). For each domain a target corpus and vocabulary (i.e. hypernym search space) are provided. The dataset contains both concepts (e.g. dog) and entities (e.g. Manchester United) up to trigrams. The following table lists the number of hyponym-hypernym pairs for each dataset:

PartitionGeneralMedicalMusic
Trial200101355
Training1177932565455
Test704841165233

The results for each model and dataset (general, medical and music) are presented below (MFH stands for “Most Frequent Hypernyms” and is used as a baseline).

General:

ModelMAPMRRP@5Paper / Source
CRIM (Bernier-Colborne and Barrière, 2018)19.7836.1019.03A Hybrid Approach to Hypernym Discovery
vTE (Espinosa-Anke et al., 2016)10.6023.839.91Supervised Distributional Hypernym Discovery via Domain Adaptation
NLP_HZ (Qui et al., 2018)9.3717.299.19A Nearest Neighbor Approach
300-sparsans (Berend et al., 2018)8.9519.448.63Hypernymy as interaction of sparse attributes
MFH8.7721.397.81--
SJTU BCMI (Zhang et al., 2018)5.7710.565.96Neural Hypernym Discovery with Term Embeddings
Apollo (Onofrei et al., 2018)2.686.012.69Detecting Hypernymy Relations Using Syntactic Dependencies
balAPInc (Shwartz et al., 2017)1.363.181.30Hypernyms under Siege: Linguistically-motivated Artillery for Hypernymy Detection

Medical domain:

ModelMAPMRRP@5Paper / Source
CRIM (Bernier-Colborne and Barrière, 2018)34.0554.6436.77A Hybrid Approach to Hypernym Discovery
MFH28.9335.8034.20--
300-sparsans (Berend et al., 2018)20.7540.6021.43Hypernymy as interaction of sparse attributes
vTE (Espinosa-Anke et al., 2016)18.8441.0720.71Supervised Distributional Hypernym Discovery via Domain Adaptation
EXPR (Issa Alaa Aldine et al., 2018)13.7740.7612.76A Combined Approach for Hypernym Discovery
SJTU BCMI (Zhang et al., 2018)11.6925.9511.69Neural Hypernym Discovery with Term Embeddings
ADAPT (Maldonado and Klubička, 2018)8.1320.568.32Skip-Gram Word Embeddings for Unsupervised Hypernym Discovery in Specialised Corpora
balAPInc (Shwartz et al., 2017)0.912.101.08Hypernyms under Siege: Linguistically-motivated Artillery for Hypernymy Detection

Music domain:

ModelMAPMRRP@5Paper / Source
CRIM (Bernier-Colborne and Barrière, 2018)40.9760.9341.31A Hybrid Approach to Hypernym Discovery
MFH33.3251.4835.76--
300-sparsans (Berend et al., 2018)29.5446.4328.86Hypernymy as interaction of sparse attributes
vTE (Espinosa-Anke et al., 2016)12.9939.3612.41Supervised Distributional Hypernym Discovery via Domain Adaptation
SJTU BCMI (Zhang et al., 2018)4.719.154.91Neural Hypernym Discovery with Term Embeddings
ADAPT (Maldonado and Klubička, 2018)2.637.462.64Skip-Gram Word Embeddings for Unsupervised Hypernym Discovery in Specialised Corpora
balAPInc (Shwartz et al., 2017)1.955.012.15Hypernyms under Siege: Linguistically-motivated Artillery for Hypernymy Detection

Taxonomy Enrichment

Given words that are not included in a taxonomy, the task is to associate each query word with its appropriate hypernyms. For instance, given an input word “milk” we need to provide a list of the most probable hypernyms the word could be attached to, e.g. “dairy product”, “beverage”. A word may have multiple hypernyms.

Datasets

SemEval 2016 Task 14

The SemEval-2016 Task 14 aims to enrich the WordNet taxonomy with new words and word senses. A system's task is to identify the WordNet synset with which the new word sense should be merged (i.e., the term is synonymous with those in the synset) or added as a hyponym (i.e., the new word sense is a specialization of an exisiting word sense).

The following table gives examples of word senses that are not defined in WordNet and their corresponding operations, illustrating the type of data that might be seen in the task.

OOV wordDefinitionTarget synsetOperation
geoscience#nany of several sciences that deal with the Earthearth_science (any of the sciences that deal with the earth or its parts)MERGE
mudslide#na mixed drink consisting of vodka, Kahlua and Bailey's.cocktail (a short mixed drink)ATTACH
euthanize#vto submit or animal to euthanasiadestroy, put down (put (an animal) to death)MERGE

The SemEval-2016 taxonomy enrichment evaluation benchmark (Jurgens and Pilehvar 2016), which can be freely downloaded here.

Novel concepts were limited to nouns and verbs, as only these parts of speech have fully-developed taxonomies in WordNet. For each item, in addition to the target synset and the operation MERGE/ATTACH, the glosses were also provided along with the source URL from which the new word sense was obtained. The dataset consists of a total of 1000 items, split into training and test datasets containing 400 and 600 items, respectively. The following table lists the number of items for each dataset:

PartitionNounVerb
Trial9334
Training34951
Test51684

The results for each model participant are presented below.

ModelLemma MatchWu&PRecallF1Paper / Source
MSejrKU (Schlichtkrull and Alonso, 2016)0.4280.5230.9730.680MSejrKu at SemEval-2016 Task 14: Taxonomy Enrichment by Evidence Ranking
TALN (Anke et al., 2016)0.3600.4761.0000.645Semantic Taxonomy Enrichment Via Sense-Based Embeddings
VCU (McInnes, 2016)0.1610.4320.9970.602Evaluating definitional-based similarity measure for semantic taxonomy enrichment
Duluth (Pedersen, 2016)0.0430.3471.0000.515Extending Gloss Overlaps to Enrich Semantic Taxonomies
Deftor (Tanev and Rotondi, 2016)0.0660.3470.9870.513Taxonomy Enrichment using Definition Vectors
UMNDuluth (Rusert and Pedersen, 2016)0.0980.3400.9980.507WordNet’s Missing Lemmas
Baseline: First word, first sense (Jurgens and Pilehvar, 2016)0.4150.5141.0000.679SemEval-2016 Task 14: Semantic Taxonomy Enrichment
Baseline: Random synset (Jurgens and Pilehvar, 2016)0.0000.2271.0000.370SemEval-2016 Task 14: Semantic Taxonomy Enrichment

Diachronic WordNet Datasets

The SemEval-2016 Task 14 setting implies pre-defined glosses. However, it is possible that new words that should be added to the taxonomy may have no definition in any other sources: they could be very rare (“apparatchik”, “falanga”), relatively new (“selfie”, “hashtag”) or come from a narrow domain (“vermiculite”).

Nikishina et al., 2020 created multiple datasets for studying diachronic evolution of wordnets, which can be downloaded from here. They chose two versions of WordNet and then select words which appear only in a newer version. For each word, they got its hypernyms from the newer WordNet version and consider them as gold standard hypernyms. The words were added to the dataset if only their hypernyms appear in both snippets. They skipped one or more WordNet versions, otherwise the dataset would be too small.

DatasetNounVerb
WordNet 1.6 - WordNet 3.017 043755
WordNet 1.7 - WordNet 3.06 161362
WordNet 2.0 - WordNet 3.02 620193

The results for each system on the current dataset are presented below.

WordNet 1.6 - WordNet 3.0
ModelMAP (Nouns)MAP (Verbs)Paper / Source
DWRank-Meta (Meta-embeddings based on Word and Graph Embeddings)0.3670.288Taxonomy enrichment with text and graph vector representations
AAEME triplet loss (Tikhomirov and Loukachevitch, 2021)0.3450.289Meta-Embeddings in Taxonomy Enrichment Task
Ranking + Wiki (Nikishina et al., 2020)0.3370.270Studying Taxonomy Enrichment on Diachronic WordNet Versions
Ranking + Wiki + node2vec + Poincare (Nikishina et al., 2021)0.3110.251Exploring Graph-based Representations for Taxonomy Enrichment
WordNet 1.7 - WordNet 3.0
ModelMAP (Nouns)MAP (Verbs)Paper / Source
DWRank-Meta (Meta-embeddings based on Word and Graph Embeddings)0.4180.227Taxonomy enrichment with text and graph vector representations
AAEME triplet loss (Tikhomirov and Loukachevitch, 2021)0.3940.239Meta-Embeddings in Taxonomy Enrichment Task
Ranking + Wiki (Nikishina et al., 2020)0.3800.200Studying Taxonomy Enrichment on Diachronic WordNet Versions
Ranking + Wiki + node2vec + Poincare (Nikishina et al., 2021)0.3500.177Exploring Graph-based Representations for Taxonomy Enrichment
WordNet 2.0 - WordNet 3.0
ModelMAP (Nouns)MAP (Verbs)Paper / Source
DWRank-Meta (Meta-embeddings based on Word and Graph Embeddings)0.4800.280Taxonomy enrichment with text and graph vector representations
AAEME triplet loss (Tikhomirov and Loukachevitch, 2021)0.4450.272Meta-Embeddings in Taxonomy Enrichment Task
Ranking + Wiki (Nikishina et al., 2020)0.4000.238Studying Taxonomy Enrichment on Diachronic WordNet Versions
Ranking + Wiki + node2vec + Poincare (Nikishina et al., 2021)0.3000.248Exploring Graph-based Representations for Taxonomy Enrichment