Back to Nlp Progress

Relationship Extraction

english/relationship_extraction.md

0.319.9 KB
Original Source

Relationship Extraction

Relationship extraction is the task of extracting semantic relationships from a text. Extracted relationships usually occur between two or more entities of a certain type (e.g. Person, Organisation, Location) and fall into a number of semantic categories (e.g. married to, employed by, lives in).

Capturing discriminative attributes (SemEval 2018 Task 10)

Capturing discriminative attributes (SemEval 2018 Task 10) is a binary classification task where participants were asked to identify whether an attribute could help discriminate between two concepts. Unlike other word similarity prediction tasks, this task focuses on the semantic differences between words.

e.g. red(attribute) can be used to discriminate apple (concept1) from banana (concept2) -> label 1

More examples:

Attributeconcept1concept2label
bookcasefridgewood1
bucketmuground0
anglecurvesharp1
pelicanturtlewater0
wirecoilmetal0

Task paper: https://www.aclweb.org/anthology/S18-1117

Task Codalab: https://competitions.codalab.org/competitions/17326

ModelExplainabilityF1 ScorePaper / SourceCode
SVM with GloVeNone0.76SUNNYNLP at SemEval-2018 Task 10: A Support-Vector-Machine-Based Method for Detecting Semantic Difference using Taxonomy and Word Embedding FeaturesAuthor's
SVM with ConceptNet, Wikipedia articles and WordNet synonymsNone0.74Luminoso at SemEval-2018 Task 10: Distinguishing Attributes Using Text Corpora and Relational KnowledgeAuthor's
MLP combining information from various DSMs, PMI, and ConceptNetNone0.73THU NGN at SemEval-2018 Task 10: Capturing Discriminative Attributes with MLP-CNN model
Gradient boosting with co-occurrence count features and JoBimText featuresNone0.73BomJi at SemEval-2018 Task 10: Combining Vector-, Pattern- and Graph-based Information to Identify Discriminative Attributes
LexVec, word co-occurrence, and ConceptNet data combined using maximum entropy classifierNone0.72UWB at SemEval-2018 Task 10: Capturing Discriminative Attributes from Word DistributionsAuthor's
Composes explicit vector spaces from WordNet Definitions, ConceptNet and Visual GenomeFully Explainable0.69Identifying and Explaining Discriminative AttributesAuthor's
Word2Vec cosine similarities of WordNet glosses Transp. (No expl.)Transp. (No expl.)0.69Meaning space at SemEval-2018 Task 10: Combining explicitly encoded knowledge with information extracted from word embeddingsAuthor's
Use of Wikipedia and ConceptNet Transp. (No expl.)Transp. (No expl.)0.69ELiRF-UPV at SemEval-2018 Task 10: Capturing Discriminative Attributes with Knowledge Graphs and Wikipedia

FewRel

The Few-Shot Relation Classification Dataset (FewRel) is a different setting from the previous datasets. This dataset consists of 70K sentences expressing 100 relations annotated by crowdworkers on Wikipedia corpus. The few-shot learning task follows the N-way K-shot meta learning setting.

The public leaderboard is available on the FewRel website.

FewRel 2

FewRel 2 extends FewRel on (1) Adaptibility to a new domain with only a hand-ful of instances (2) Ability to detect none-of-the-above relations? The paper is at ACL Web.

The public leaderboard is available on FewRel 2 website

Multi-Way Classification of Semantic Relations Between Pairs of Nominals (SemEval 2010 Task 8)

SemEval-2010 introduced 'Task 8 - Multi-Way Classification of Semantic Relations Between Pairs of Nominals'. The task is, given a sentence and two tagged nominals, to predict the relation between those nominals and the direction of the relation. The dataset contains nine general semantic relations together with a tenth 'OTHER' relation.

Example:

There were apples, pears and oranges in the bowl.

(content-container, pears, bowl)

The main evaluation metric used is macro-averaged F1, averaged across the nine proper relationships (i.e. excluding the OTHER relation), taking directionality of the relation into account.

Several papers have used additional data (e.g. pre-trained word embeddings, WordNet) to improve performance. The figures reported here are the highest achieved by the model using any external resources.

End-to-End Models

ModelF1Paper / SourceCode
BERT-based Models
A-GCN (Tian et al., 2021)89.85Dependency-driven Relation Extraction with Attentive Graph Convolutional NetworksOfficial
Matching-the-Blanks (Baldini Soares et al., 2019)89.5Matching the Blanks: Distributional Similarity for Relation Learning
R-BERT (Wu et al. 2019)89.25Enriching Pre-trained Language Model with Entity Information for Relation Classificationmickeystroller's Reimplementation
CNN-based Models
Multi-Attention CNN (Wang et al. 2016)88.0Relation Classification via Multi-Level Attention CNNslawlietAi's Reimplementation
Attention CNN (Huang and Y Shen, 2016)84.3
85.9<sup>*</sup>Attention-Based Convolutional Neural Network for Semantic Relation Extraction
CR-CNN (dos Santos et al., 2015)84.1Classifying Relations by Ranking with Convolutional Neural Networkpratapbhanu's Reimplementation
CNN (Zeng et al., 2014)82.7Relation Classification via Convolutional Deep Neural Networkroomylee's Reimplementation
RNN-based Models
Entity Attention Bi-LSTM (Lee et al., 2019)85.2Semantic Relation Classification via Bidirectional LSTM Networks with Entity-aware Attention using Latent Entity TypingOfficial
Hierarchical Attention Bi-LSTM (Xiao and C Liu, 2016)84.3Semantic Relation Classification via Hierarchical Recurrent Neural Network with Attention
Attention Bi-LSTM (Zhou et al., 2016)84.0Attention-Based Bidirectional Long Short-Term Memory Networks for Relation ClassificationSeoSangwoo's Reimplementation
Bi-LSTM (Zhang et al., 2015)82.7
84.3<sup>*</sup>Bidirectional long short-term memory networks for relation classification

<a name="footnote">*</a>: It uses external lexical resources, such as WordNet, part-of-speech tags, dependency tags, and named entity tags.

Dependency Models

ModelF1Paper / SourceCode
BRCNN (Cai et al., 2016)86.3Bidirectional Recurrent Convolutional Neural Network for Relation Classification
DRNNs (Xu et al., 2016)86.1Improved Relation Classification by Deep Recurrent Neural Networks with Data Augmentation
depLCNN + NS (Xu et al., 2015a)85.6Semantic Relation Classification via Convolutional Neural Networks with Simple Negative Sampling
SDP-LSTM (Xu et al., 2015b)83.7Classifying Relations via Long Short Term Memory Networks along Shortest Dependency PathSshanu's Reimplementation
DepNN (Liu et al., 2015)83.6A Dependency-Based Neural Network for Relation Classification
FCN (Yu et al., 2014)83.0Factor-based compositional embedding models
MVRNN (Socher et al., 2012)82.4Semantic Compositionality through Recursive Matrix-Vector Spacespratapbhanu's Reimplementation

New York Times Corpus

The standard corpus for distantly supervised relationship extraction is the New York Times (NYT) corpus, published in Riedel et al, 2010.

This contains text from the New York Times Annotated Corpus with named entities extracted from the text using the Stanford NER system and automatically linked to entities in the Freebase knowledge base. Pairs of named entities are labelled with relationship types by aligning them against facts in the Freebase knowledge base. (The process of using a separate database to provide label is known as 'distant supervision')

Example:

Elevation Partners, the $1.9 billion private equity group that was founded by Roger McNamee

(founded_by, Elevation_Partners, Roger_McNamee)

Different papers have reported various metrics since the release of the dataset, making it difficult to compare systems directly. The main metrics used are either precision at N results or plots of the precision-recall. The range of recall has increased over the years as systems improve, with earlier systems having very low precision at 30% recall.

ModelP@10%P@30%Paper / SourceCode
KGPOOL (Nadgeri et al., 2021)92.386.7KGPool: Dynamic Knowledge Graph Context Selection for Relation ExtractionKGPOOL
RECON (Bastos et al., 2021)87.574.1RECON: Relation Extraction using Knowledge Graph Context in a Graph Neural NetworkRECON
HRERE (Xu et al., 2019)84.972.8Connecting Language and Knowledge with Heterogeneous Representations for Neural Relation ExtractionHRERE
PCNN+noise_convert+cond_opt (Wu et al., 2019)81.761.8Improving Distantly Supervised Relation Extraction with Neural Noise Converter and Conditional Optimal Selector
Intra- and Inter-Bag (Ye and Ling, 2019)78.962.4Distant Supervision Relation Extraction with Intra-Bag and Inter-Bag AttentionsCode
RESIDE (Vashishth et al., 2018)73.659.5RESIDE: Improving Distantly-Supervised Neural Relation Extraction using Side InformationRESIDE
PCNN+ATT (Lin et al., 2016)69.451.8Neural Relation Extraction with Selective Attention over InstancesOpenNRE
MIML-RE (Surdeneau et al., 2012)60.7+-Multi-instance Multi-label Learning for Relation ExtractionMimlre
MultiR (Hoffman et al., 2011)60.9+-Knowledge-Based Weak Supervision for Information Extraction of Overlapping RelationsMultiR
(Mintz et al., 2009)39.9+-Distant supervision for relation extraction without labeled data

(+) Obtained from results in the paper "Neural Relation Extraction with Selective Attention over Instances"

WikiData dataset for Sentential Relation Extraction

The sentential RE ignores any other occurrence of the given entity pair, thereby making the target relation predictions on the sentence level (Sorokin and Gurevych, 2017). The paper introduces a dataset on Wikidata KG containing 353 relations.

ModelF1Paper / SourceCode
KGPOOL (Nadgeri et al., 2021)88.60KGPool: Dynamic Knowledge Graph Context Selection for Relation Extraction
RECON (Bastos et al., 2021)87.23RECON: Relation Extraction using Knowledge Graph Context in a Graph Neural Network
GPGNN (Zhu et al., 2019)82.29Graph Neural Networks with Generated Parameters for Relation Extraction
ContextAware (Sorokin and Gurevych, 2017)72.07Context-Aware Representations for Knowledge Base Relation Extraction

Joint Entity and Relation Extraction

In this task binary relation tuples (two entities and a relation between them) are jointly extracted from sentences. The input to the models is just the sentences and a set of relations, output is a set of relation tuples. Models should extract all relation tuples present in the sentences with full entity names and overlapping entities. F1 score is used to evaluate the models. An extracted tuple is considered as correct if the two entities and the relation match with a ground truth tuple.

NYT29

This dataset is derived from the New York Times dataset of Riedel et al., 2010. It has 29 relations.

ModelF1Paper / SourceCode
WDec (Nayak and Ng, 2020)0.682Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation ExtractionPtrNetDecoding4JERE
PNDec (Nayak and Ng, 2020)0.673Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation ExtractionPtrNetDecoding4JERE
HRLRE (Takanobu et at., 2019)0.643A Hierarchical Framework for Relation Extraction with Reinforcement LearningHRLRE
NYT24

This dataset is derived from the New York Times dataset of Hoffman et al., 2011. It has 24 relations.

ModelF1Paper / SourceCode
WDec (Nayak and Ng, 2020)0.817Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation ExtractionPtrNetDecoding4JERE
PNDec (Nayak and Ng, 2020)0.789Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation ExtractionPtrNetDecoding4JERE
HRLRE (Takanobu et at., 2019)0.776A Hierarchical Framework for Relation Extraction with Reinforcement LearningHRLRE

TACRED

TACRED is a large-scale relation extraction dataset with 106,264 examples built over newswire and web text from the corpus used in the yearly TAC Knowledge Base Population (TAC KBP) challenges. Examples in TACRED cover 41 relation types as used in the TAC KBP challenges (e.g., per:schools_attended and org:members) or are labeled as no_relation if no defined relation is held. These examples are created by combining available human annotations from the TAC KBP challenges and crowdsourcing.

Example:

Billy Mays, the bearded, boisterious pitchman who, as the undisputed king of TV yell and sell, became an inlikely pop culture icon, died at his home in Tampa, Fla, on Sunday.

(per:city_of_death, Billy Mays, Tampa)

The main evaluation metric used is micro-averaged F1 over instances with proper relationships (i.e. excluding the no_relation type).

ModelF1Paper / SourceCode
LUKE (Yamada et al., 2020)72.7LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attentionOfficial
Matching-the-Blanks (Baldini Soares et al., 2019)71.5Matching the Blanks: Distributional Similarity for Relation Learning
C-GCN + PA-LSTM (Zhang et al. 2018)68.2Graph Convolution over Pruned Dependency Trees Improves Relation ExtractionOffical
PA-LSTM (Zhang et al, 2017)65.1Position-aware Attention and Supervised Data Improve Slot FillingOfficial

Go back to the README