Back to Nlp Progress

Dialogue

english/dialogue.md

0.325.7 KB
Original Source

Dialogue

Dialogue is notoriously hard to evaluate. Past approaches have used human evaluation.

Dialogue act classification

Dialogue act classification is the task of classifying an utterance with respect to the function it serves in a dialogue, i.e. the act the speaker is performing. Dialogue acts are a type of speech acts (for Speech Act Theory, see Austin (1975) and Searle (1969)).

Switchboard corpus

The Switchboard-1 corpus is a telephone speech corpus, consisting of about 2,400 two-sided telephone conversation among 543 speakers with about 70 provided conversation topics. The dataset includes the audio files and the transcription files, as well as information about the speakers and the calls.

The Switchboard Dialogue Act Corpus (SwDA) [download] extends the Switchboard-1 corpus with tags from the SWBD-DAMSL tagset, which is an augmentation to the Discourse Annotation and Markup System of Labeling (DAMSL) tagset. The 220 tags were reduced to 42 tags by clustering in order to improve the language model on the Switchboard corpus. A subset of the Switchboard-1 corpus consisting of 1155 conversations was used. The resulting tags include dialogue acts like statement-non-opinion, acknowledge, statement-opinion, agree/accept, etc.
Annotated example:
Speaker: A, Dialogue Act: Yes-No-Question, Utterance: So do you go to college right now?

ModelAccuracyPaper / SourceCode
SGNN (Ravi et al., 2018)83.1Self-Governing Neural Networks for On-Device Short Text ClassificationLink
CASA (Raheja et al., 2019)82.9Dialogue Act Classification with Context-Aware Self-AttentionLink
DAH-CRF (Li et al., 2019)82.3A Dual-Attention Hierarchical Recurrent Neural Network for Dialogue Act Classification
ALDMN (Wan et al., 2018)81.5Improved Dynamic Memory Network for Dialogue Act Classification with Adversarial Training
CRF-ASN (Chen et al., 2018)81.3Dialogue Act Recognition via CRF-Attentive Structured Network
Bi-LSTM-CRF (Kumar et al., 2017)79.2Dialogue Act Sequence Labeling using Hierarchical encoder with CRFLink
RNN with 3 utterances in context (Bothe et al., 2018)77.34A Context-based Approach for Dialogue Act Recognition using Simple Recurrent Neural Networks

ICSI Meeting Recorder Dialog Act (MRDA) corpus

The MRDA corpus [download] consists of about 75 hours of speech from 75 naturally-occurring meetings among 53 speakers. The tagset used for labeling is a modified version of the SWBD-DAMSL tagset. It is annotated with three types of information: marking of the dialogue act segment boundaries, marking of the dialogue acts and marking of correspondences between dialogue acts.
Annotated example:
Time: 2804-2810, Speaker: c6, Dialogue Act: s^bd, Transcript: i mean these are just discriminative.
Multiple dialogue acts are separated by "^".

ModelAccuracyPaper / SourceCode
DAH-CRF (Li et al., 2019)92.2A Dual-Attention Hierarchical Recurrent Neural Network for Dialogue Act Classification
CRF-ASN (Chen et al., 2018)91.7Dialogue Act Recognition via CRF-Attentive Structured Network
CASA (Raheja et al., 2019)91.1Dialogue Act Classification with Context-Aware Self-Attention
Bi-LSTM-CRF (Kumar et al., 2017)90.9Dialogue Act Sequence Labeling using Hierarchical encoder with CRFLink
SGNN (Ravi et al., 2018)86.7Self-Governing Neural Networks for On-Device Short Text Classification

Dialogue state tracking

Dialogue state tacking consists of determining at each turn of a dialogue the full representation of what the user wants at that point in the dialogue, which contains a goal constraint, a set of requested slots, and the user's dialogue act.

Second dialogue state tracking challenge

For goal-oriented dialogue, the dataset of the second Dialogue Systems Technology Challenges (DSTC2) is a common evaluation dataset. The DSTC2 focuses on the restaurant search domain. Models are evaluated based on accuracy on both individual and joint slot tracking.

ModelRequestAreaFoodPriceJointPaper / Source
Zhong et al. (2018)97.5---74.5Global-locally Self-attentive Dialogue State Tracker
Liu et al. (2018)-90849272Dialogue Learning with Human Teaching and Feedback in End-to-End Trainable Task-Oriented Dialogue Systems
Neural belief tracker (Mrkšić et al., 2017)96.590849473.4Neural Belief Tracker: Data-Driven Dialogue State Tracking
RNN (Henderson et al., 2014)95.792868669Robust dialog state tracking using delexicalised recurrent neural networks and unsupervised gate

Wizard-of-Oz

The WoZ 2.0 dataset is a newer dialogue state tracking dataset whose evaluation is detached from the noisy output of speech recognition systems. Similar to DSTC2, it covers the restaurant search domain and has identical evaluation.

ModelRequestJointPaper / Source
BERT-based tracker (Lai et al., 2020)97.690.5A Simple but Effective BERT Model for Dialog State Tracking on Resource-Limited Systems
GCE (Nouri et al., 2018)97.488.5Toward Scalable Neural Dialogue State Tracking Model
Zhong et al. (2018)97.188.1Global-locally Self-attentive Dialogue State Tracker
Neural belief tracker (Mrkšić et al., 2017)96.584.4Neural Belief Tracker: Data-Driven Dialogue State Tracking
RNN (Henderson et al., 2014)87.170.8Robust dialog state tracking using delexicalised recurrent neural networks and unsupervised gate

MultiWOZ

The MultiWOZ dataset is a fully-labeled collection of human-human written conversations spanning over multiple domains and topics. At a size of 10k dialogues, it is at least one order of magnitude larger than all previous annotated task-oriented corpora. The dialogue are set between a tourist and a clerk in the information. It spans over 7 domains.

Belief Tracking

<div class="datagrid" style="width:500px;"> <table> <thead><tr><th></th><th colspan="2">MultiWOZ 2.0</th><th colspan="2">MultiWOZ 2.1</th></tr></thead> <thead><tr><th>Model</th><th>Joint Accuracy</th><th>Slot</th><th>Joint Accuracy</th><th>Slot</th></tr></thead> <tbody> <tr><td><a href="https://www.aclweb.org/anthology/P18-2069">MDBT</a> (Ramadan et al., 2018) </td><td>15.57 </td><td>89.53</td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/abs/1805.09655">GLAD</a> (Zhong et al., 2018)</td><td>35.57</td><td>95.44 </td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1812.00899.pdf">GCE</a> (Nouri and Hosseini-Asl, 2018)</td><td>36.27</td><td>98.42</td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1908.01946.pdf">Neural Reading</a> (Gao et al, 2019)</td><td>41.10</td><td></td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1907.00883.pdf">HyST</a> (Goel et al, 2019)</td><td>44.24</td><td></td><td></td><td></td></tr> <tr><td><a href="https://www.aclweb.org/anthology/P19-1546/">SUMBT</a> (Lee et al, 2019)</td><td>46.65</td><td>96.44</td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1905.08743.pdf">TRADE</a> (Wu et al, 2019)</td><td>48.62</td><td>96.92</td><td>45.60</td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1909.00754.pdf">COMER</a> (Ren et al, 2019)</td><td>48.79</td><td></td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/1911.06192.pdf">DSTQA</a> (Zhou et al, 2019)</td><td>51.44</td><td>97.24</td><td>51.17</td><td>97.21</td></tr> <tr><td><a href="https://arxiv.org/pdf/1910.03544.pdf">DST-Picklist</a> (Zhang et al, 2019)</td><td></td><td></td><td>53.3</td><td></td></tr> <tr><td><a href="https://www.aaai.org/Papers/AAAI/2020GB/AAAI-ChenL.10030.pdf">SST</a> (Chen et al. 2020)</td><td></td><td></td><td>55.23</td><td></td></tr> <tr><td><a href="https://arxiv.org/abs/2005.02877">TripPy</a> (Heck et al. 2020)</td><td></td><td></td><td>55.3</td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/2005.00796.pdf">SimpleTOD</a> (Hosseini-Asl et al. 2020)</td><td></td><td></td><td>55.72</td><td></td></tr> </tbody> </table> </div>

Policy Optimization

<div class="datagrid" style="width:500px;"> <table> <thead><tr><th>(INFORM + SUCCESS)*0.5 + BLEU</th><th colspan="3">MultiWOZ 2.0</th><th colspan="3">MultiWOZ 2.1</th></tr></thead> <thead><tr><th>Model</th><th>INFORM</th><th>SUCCESS</th><th>BLEU</th><th>INFORM</th><th>SUCCESS</th><th>BLEU</th></tr></thead> <tbody> <tr><td><a href="https://arxiv.org/pdf/1907.05346.pdf">TokenMoE</a> (Pei et al. 2019)</td><td>75.30</td><td> 59.70</td><td> 16.81 </td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://pdfs.semanticscholar.org/47d0/1eb59cd37d16201fcae964bd1d2b49cfb55e.pdf">Baseline</a> (Budzianowski et al. 2018)</td><td>71.29</td><td> 60.96 </td><td> 18.8 </td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/1907.10016.pdf">Structured Fusion</a> (Mehri et al. 2019)</td><td>82.70</td><td>72.10</td><td> 16.34</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/abs/1902.08858">LaRL</a> (Zhao et al. 2019)</td><td>82.8</td><td>79.2</td><td> 12.8</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/2005.00796.pdf">SimpleTOD</a> (Hosseini-Asl et al. 2020)</td><td>88.9</td><td>67.1</td><td> 16.9</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/1911.08151.pdf">MoGNet</a> (Pei et al. 2019)</td><td>85.3</td><td>73.30</td><td> 20.13</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/1905.12866.pdf">HDSA</a> (Chen et al. 2019)</td><td>82.9</td><td>68.9</td><td> 23.6</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/abs/1910.03756">ARDM</a> (Wu et al. 2019)</td><td>87.4</td><td>72.8</td><td> 20.6</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/1911.10484.pdf">DAMD</a> (Zhang et al. 2019)</td><td>89.2</td><td>77.9</td><td> 18.6</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/2005.05298.pdf">SOLOIST</a> (Peng et al. 2020)</td><td>89.60</td><td> 79.30</td><td> 18.3</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/2004.12363.pdf">MarCo</a> (Wang et al. 2020)</td><td>92.30</td><td> 78.60</td><td> 20.02</td><td> 92.50</td><td> 77.80</td><td> 19.54</td></tr> <tfoot> </tfoot> </tbody> </table> </div>

Natural Language Generation

<div class="datagrid" style="width:500px;"><table> <thead><tr><th>Model</th><th>SER</th><th>BLEU</th></tr></thead> <tbody> <tr><td><a href="https://pdfs.semanticscholar.org/47d0/1eb59cd37d16201fcae964bd1d2b49cfb55e.pdf">Baseline</a> (Budzianowski et al. 2018)</td><td>2.99 </td><td> 0.632</td></tr> </tbody> </table> </div>

End-to-End Modelling

<div class="datagrid" style="width:500px;"> <table> <thead><tr><th>(INFORM + SUCCESS)*0.5 + BLEU</th><th colspan="3">MultiWOZ 2.0</th><th colspan="3">MultiWOZ 2.1</th></tr></thead> <thead><tr><th>Model</th><th>INFORM</th><th>SUCCESS</th><th>BLEU</th><th>INFORM</th><th>SUCCESS</th><th>BLEU</th></tr></thead> <tbody> <tr><td><a href="https://arxiv.org/pdf/1911.10484.pdf">DAMD</a> (Zhang et al. 2019)</td><td>76.3</td><td>60.4</td><td> 18.6</td><td> </td><td> </td><td> </td></tr> <tr><td><a href="https://arxiv.org/pdf/2005.00796.pdf">SimpleTOD</a> (Hosseini-Asl et al. 2020)</td><td>84.4</td><td>70.1</td><td> 15.01</td><td> </td><td></td><td></td></tr> <tr><td><a href="https://arxiv.org/pdf/2005.05298.pdf">SOLOIST</a> (Peng et al. 2020)</td><td>85.50</td><td>72.90</td><td> 16.54</td><td> </td><td></td><td> </td></tr> <tfoot> </tfoot> </tbody> </table> </div>

Retrieval-based Chatbots

These systems take as input a context and a list of possible responses and rank the responses, returning the highest ranking one.

Ubuntu IRC Data

There are several corpra based on the Ubuntu IRC Channel Logs:

Each version of the dataset contains a set of dialogues from the IRC channel, extracted by automatically disentangling conversations occurring simultaneously. See below for results on the disentanglement process.

The exact tasks used vary slightly, but all consider variations of Recall_N@K, which means how often the true answer is in the top K options when there are N total candidates.

DataModelR_100@1R_100@10R_100@50MRRPaper / Source
DSTC 8 (main)Wu et. al., (2020)76.197.9-84.8Enhancing Response Selection with Advanced Context Modeling and Post-training
DSTC 8 (subtask 2)Wu et. al., (2020)70.695.7-79.9Enhancing Response Selection with Advanced Context Modeling and Post-training
DSTC 7Seq-Att-Network (Chen and Wang, 2019)64.590.299.473.5Sequential Attention-based Network for Noetic End-to-End Response Selection
DataModelR_2@1R_10@1Paper / Source
UDC v2DAM (Zhou et al. 2018)93.876.7Multi-Turn Response Selection for Chatbots with Deep Attention Matching Network
UDC v2SMN (Wu et al. 2017)92.372.3Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots
UDC v2Multi-View (Zhou et al. 2017)90.866.2Multi-view Response Selection for Human-Computer Conversation
UDC v2Bi-LSTM (Kadlec et al. 2015)89.563.0Improved Deep Learning Baselines for Ubuntu Corpus Dialogs

Additional results can be found in the DSTC task reports linked above.

Reddit Corpus

The Reddit Corpus contains 726 million multi-turn dialogues from the Reddit board. Reddit is an American social news aggregation website, where users can post links, and take partin discussions on these post. The task of Reddit Corpus is to select the correct response from 100 candidates (others are negatively sampled) by considering previous conversation history. Models are evaluated with the Recall 1 at 100 metric (the 1-of-100 ranking accuracy). You can find more details at here.

ModelR_1@100Paper / Source
PolyAI Encoder (Henderson et al. 2019)61.3A Repository of Conversational Dataset
USE (Cer et al. 2018)47.7Universal Sentence Encoder
BERT (Devlin et al. 2017)24.0BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
ELMO (Peters et al. 2018)19.3Deep contextualized word representations

Advising Corpus

The Advising Corpus, available here, contains a collection of conversations between a student and an advisor at the University of Michigan. They were released as part of DSTC 7 track 1 and used again in DSTC 8 track 2.

ModelR_100@1R_100@10R_100@50MRRPaper / Source
Yang et. al., (2020)56.487.8-67.7Transformer-based Semantic Matching Model for Noetic Response Selection
Seq-Att-Network (Chen and Wang, 2019)21.463.094.833.9Sequential Attention-based Network for Noetic End-to-End Response Selection

Generative-based Chatbots

The main task of generative-based chatbot is to generate consistent and engaging response given the context.

Personalized Chit-chat

The task of persinalized chit-chat dialogue generation is first proposed by PersonaChat. The motivation is to enhance the engagingness and consistency of chit-chat bots via endowing explicit personas to agents. Here the persona is defined as several profile natural language sentences like "I weight 300 pounds.". NIPS 2018 has hold a competition The Conversational Intelligence Challenge 2 (ConvAI2) based on the dataset. The Evaluation metric is F1, Hits@1 and ppl. F1 evaluates on the word-level, and Hits@1 represents the probability of the real next utterance ranking the highest according to the model, while ppl is perplexity for language modeling. The following results are reported on dev set (test set is still hidden), almost of them are borrowed from ConvAI2 Leaderboard.

ModelF1Hits@1pplPaper / SourceCode
P^2 Bot (Liu et al. 2020)19.7781.915.12You Impress Me: Dialogue Generation via Mutual Persona PerceptionCode
TransferTransfo (Thomas et al. 2019)19.0982.117.51TransferTransfo: A Transfer Learning Approach for Neural Network Based Conversational AgentsCode
Lost In Conversation17.79-17.3NIPS 2018 Workshop PresentationCode
Seq2Seq + Attention (Dzmitry et al. 2014)16.1812.629.8Neural Machine Translation by Jointly Learning to Align and TranslateCode
KV Profile Memory (Zhang et al. 2018)11.955.2-Personalizing Dialogue Agents: I have a dog, do you have pets too?Code

Disentanglement

As noted for the Ubuntu data above, sometimes multiple conversations are mixed together in a single channel. Work on conversation disentanglement aims to separate out conversations. There are two main resources for the task.

This can be formultated as a clustering problem, with no clear best metric. Several metrics are considered:

  • Variation of Information
  • F-1 over 1-1 matched clusters using max-flow
  • Precision, Recall, and F-score on exact match for clusters
  • Local overlap
  • Another form of F-1 defined by Shen et al. (2006)

Ubuntu IRC

Manually labeled by Kummerfeld et al. (2019), this data is available here.

ModelVI1-1PrecisionRecallF-ScorePaper / SourceCode
BERT + BiLSTM93.3-44.349.646.8Pre-Trained and Attention-Based Neural Networks for Building Noetic Task-Oriented Dialogue Systems-
FF ensemble: Vote (Kummerfeld et al., 2019)91.576.036.339.738.0A Large-Scale Corpus for Conversation DisentanglementCode
Feedforward (Kummerfeld et al., 2019)91.375.634.638.036.2A Large-Scale Corpus for Conversation DisentanglementCode
FF ensemble: Intersect (Kummerfeld et al., 2019)69.326.667.021.132.1A Large-Scale Corpus for Conversation DisentanglementCode
Linear (Elsner and Charniak, 2008)82.151.412.121.515.5You Talking to Me? A Corpus and Algorithm for Conversation DisentanglementCode
Heuristic (Lowe et al., 2015)80.653.710.87.68.9Training End-to-End Dialogue Systems with the Ubuntu Dialogue CorpusCode

Linux IRC

This data has been manually annotated three times:

DataModel1-1LocalShen F-1Paper / SourceCode
KummerfeldLinear (Elsner and Charniak, 2008)59.780.863.0You Talking to Me? A Corpus and Algorithm for Conversation DisentanglementCode
KummerfeldFeedforward (Kummerfeld et al., 2019)57.780.359.8A Large-Scale Corpus for Conversation DisentanglementCode
KummerfeldHeuristic (Lowe et al., 2015)43.467.950.7Training End-to-End Dialogue Systems with the Ubuntu Dialogue CorpusCode
ElsnerLinear (Elsner and Charniak, 2008)53.181.955.1You Talking to Me? A Corpus and Algorithm for Conversation DisentanglementCode
ElsnerFeedforward (Kummerfeld et al., 2019)52.177.853.8A Large-Scale Corpus for Conversation DisentanglementCode
ElsnerWang and Oard (2009)47.075.152.8Context-based Message Expansion for Disentanglement of Interleaved Text Conversations-
ElsnerHeuristic (Lowe et al., 2015)45.173.851.8Training End-to-End Dialogue Systems with the Ubuntu Dialogue CorpusCode