SlideShare a Scribd company logo
Character Identification on Multiparty Conversation:
Identifying Mentions of Characters in TV Shows
Henry (Yu-Hsin) Chen and Jinho D. Choi
Department of Mathematics and Computer Science, Emory University
• Define the task of “Character Identification”, a sub-task of entity linking.
• Generate and analyze dialogue-specific corpus for the task.
• Tackle the task via adaptation of existing coreference resolution systems.
• Evaluate the performance of coreference resolution systems on the task.
Objective
• Given corpora, character identification can be solved with trained models
• Existing coreference resolution systems can be modified to solve this task.
• Models trained on one domain can be used to decode on other domains.
• Models perform worse when trained with data of more than one domains.
• Models trained on scene-level documents perform better in majority of the
cases, though they failed at out-of-context character inferences.
• Similar trends are observed after remapping the resultant coreferent chains
from the systems in order to tackle character identification.
Conclusion
• Creation of a character identification specific entity linking system.
• Resolution of the linking for collective and plural mentions.
• Identification of disfluency markers and singleton mentions.
• Generation of knowledge base from entity-centric attributes of mentions.
Future Work
Data Collection
• Transcripts of TV shows, “Friends”(F) and “The Big Bang Theory”(BB), are
collected from publically available websites and divided into scenes.
Corpus Annotation
• Corpus is double-annotated then adjudicated on Amazon Turk.
Inter-Annotator Agreement
• Quality of annotations is evaluated by their raw
agreement and Cohen-Kappa scores.
Corpus Creation
Season
Episode
Scene
Utterance
Speaker
Statement(s)
Utterance text
Figure 3. Structure of the corpus. Table 1. Statistics of the corpus. Epi/Sce/Spk: count of episodes, scenes,
and speakers. UC/SC/WC: count of utterances, sentences, and words
Figure 4. Template used on Amazon Turk for corpus annotation and adjudication
Table 2. Inter-Annotator Agreement scores. F1p is a preliminary annotation trial
done without context of the +-2 scenes and dynamic inferred speaker options
• Character Identification is a sub-task of entity linking. It focuses on mapping
mentions in context to one or more characters in a knowledge base.
• The task is different from coreference resolution since the each coreferent
chain does connect to an entity and is ambiguous for collective mentions.
Task Definition
Ross I told mom and dad last night, they seemed to take it pretty well.
Monica
Oh really, so that hysterical phone call I got from a woman at sobbing 3:00 A.M., "I'll
never have grandchildren, I'll never have grandchildren." was what? A wrong number?
MonicaJack JudyRoss
Character Identification
Figure 1. Task illustration of Character Identification.
Ross I told mom and dad last night, they seemed to take it pretty well.
Monica
Oh really, so that hysterical phone call I got from a woman at sobbing 3:00 A.M., "I'll
never have grandchildren, I'll never have grandchildren." was what? A wrong number?
Coreference Resolution
Figure 2. Task illustration of Coreference Resolution.
• Linguistically-motivated rules, such as name entities, pronouns, and
personal nouns, are used to select mentions from noun phrases.
• Performance of the mention detection is manually evaluated by randomly
examining 5% of the entire corpus.
• The missed and error cases
of the mention detection are
further investigated.
Mention Detection
Table 3. Statistics of mentions found in our corpus.
NE: Name entities. PRP: pronouns. PNN: personal nouns
Table 4. Analysis on the performance of our mention detection.
P: Precision. R: Recall. F: F-1 score.
Table 1
Analogous phrases 2.06% 2
Misspelled pronouns 5.15% 5
Non-nominals 7.21% 7
Proper noun misses 9.28% 9
Interjection use of
pronouns
14.43% 14
Common noun misses 14.43% 14
27%
27% 18%
14%
10%
4%
Analogous phrases
Misspelled pronouns
Non-nominals
Proper noun misses
Interjection use of pronouns
Common noun misses
1
Figure 5. Proportions of the misses and
errors of the mention detection.
Coreference Resolution
• Stanford Multi-Sieve System is a rule-based system composed of multiple
sieves of linguistic rules. It serves as the baseline of our task.
• Stanford Entity-Centric System is a statistical system that extracts global
entity-level features and constructs feature clusters for the resolution.
• Only gold mentions are used for coreference resolutions.
• Different combinations of the corpus are used to train the statistical system.
Coreferent Chain Remapping
• The resultant coreferent chains from the coreference resolution systems are
remapped to an entity through voting based on the majority of mentions.
• The referent of each mention is determined by pre-defined rules:
1. Character-related proper noun or named entity mention refers to the
character of the proper noun or the named entity.
2. First-person pronoun or possessive pronoun mention refers to the
character of the utterance contain the mention.
3. Collective pronoun or possessive pronoun refers to Collective group.
Methodology
Table 5. Corpus data split for training(TRN), developing(DEV) and testing(TST)
• CoNLL’12 official scorer is used to evaluated immediate results of the
coreference resolution systems with MUC, B3, and CEAFm metrics.
• The remapped coreferent chains are evaluated in terms of purity scores.
This represents the performance of the systems on character identification.
Evaluation
Table 6. Evaluations of the coreference resolution systems.
Document episode/scene: each episode/scene is treated as a document.
Table 7. Evaluations character identification after remapping the coreferent chains.
FC/EC/UC: Found, expected, and unknown(%) clusters. UM: unknown(%) mentions.

More Related Content

PDF
Robust Coreference Resolution and Entity Linking on Dialogues: Character Iden...
PDF
Character Identification on Multi-party Dialogues
PDF
End-to-End Plural Coreference Resolution on TV Show Transcripts
PPTX
Coreference_Resolution in Natural language processing
PDF
SemEval 2018 Task 4: Character Identification on Multiparty Dialogues
PDF
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
PDF
Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP D...
PDF
Competence-Level Prediction and Resume & Job Description Matching Using Conte...
Robust Coreference Resolution and Entity Linking on Dialogues: Character Iden...
Character Identification on Multi-party Dialogues
End-to-End Plural Coreference Resolution on TV Show Transcripts
Coreference_Resolution in Natural language processing
SemEval 2018 Task 4: Character Identification on Multiparty Dialogues
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP D...
Competence-Level Prediction and Resume & Job Description Matching Using Conte...

More from Jinho Choi (20)

PDF
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
PDF
The Myth of Higher-Order Inference in Coreference Resolution
PDF
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
PDF
Abstract Meaning Representation
PDF
Semantic Role Labeling
PDF
CKY Parsing
PDF
CS329 - WordNet Similarities
PDF
CS329 - Lexical Relations
PDF
Automatic Knowledge Base Expansion for Dialogue Management
PDF
Attention is All You Need for AMR Parsing
PDF
Graph-to-Text Generation and its Applications to Dialogue
PDF
Real-time Coreference Resolution for Dialogue Understanding
PDF
Topological Sort
PDF
Tries - Put
PDF
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
PDF
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
PDF
How to make Emora talk about Sports Intelligently
PDF
Text-to-SQL with Data-Driven Templates
PDF
Resume Classification with Term Attention Embeddings
PDF
[DSA-Java] Heap Sort
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
The Myth of Higher-Order Inference in Coreference Resolution
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
Abstract Meaning Representation
Semantic Role Labeling
CKY Parsing
CS329 - WordNet Similarities
CS329 - Lexical Relations
Automatic Knowledge Base Expansion for Dialogue Management
Attention is All You Need for AMR Parsing
Graph-to-Text Generation and its Applications to Dialogue
Real-time Coreference Resolution for Dialogue Understanding
Topological Sort
Tries - Put
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
How to make Emora talk about Sports Intelligently
Text-to-SQL with Data-Driven Templates
Resume Classification with Term Attention Embeddings
[DSA-Java] Heap Sort
Ad

Recently uploaded (20)

PDF
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PPTX
1. Introduction to Computer Programming.pptx
PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PDF
NewMind AI Weekly Chronicles - August'25-Week II
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
Univ-Connecticut-ChatGPT-Presentaion.pdf
PDF
DASA ADMISSION 2024_FirstRound_FirstRank_LastRank.pdf
PPTX
cloud_computing_Infrastucture_as_cloud_p
PDF
Encapsulation theory and applications.pdf
PDF
Mushroom cultivation and it's methods.pdf
PDF
Assigned Numbers - 2025 - Bluetooth® Document
PDF
Heart disease approach using modified random forest and particle swarm optimi...
PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PDF
Microsoft Solutions Partner Drive Digital Transformation with D365.pdf
PPTX
A Presentation on Artificial Intelligence
PDF
Web App vs Mobile App What Should You Build First.pdf
PDF
Hybrid model detection and classification of lung cancer
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
Hindi spoken digit analysis for native and non-native speakers
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
Building Integrated photovoltaic BIPV_UPV.pdf
1. Introduction to Computer Programming.pptx
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
NewMind AI Weekly Chronicles - August'25-Week II
MIND Revenue Release Quarter 2 2025 Press Release
Univ-Connecticut-ChatGPT-Presentaion.pdf
DASA ADMISSION 2024_FirstRound_FirstRank_LastRank.pdf
cloud_computing_Infrastucture_as_cloud_p
Encapsulation theory and applications.pdf
Mushroom cultivation and it's methods.pdf
Assigned Numbers - 2025 - Bluetooth® Document
Heart disease approach using modified random forest and particle swarm optimi...
Accuracy of neural networks in brain wave diagnosis of schizophrenia
Microsoft Solutions Partner Drive Digital Transformation with D365.pdf
A Presentation on Artificial Intelligence
Web App vs Mobile App What Should You Build First.pdf
Hybrid model detection and classification of lung cancer
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
Hindi spoken digit analysis for native and non-native speakers
Ad

Character Identification on Multiparty Conversation: Identifying Mentions of Characters in TV Shows

  • 1. Character Identification on Multiparty Conversation: Identifying Mentions of Characters in TV Shows Henry (Yu-Hsin) Chen and Jinho D. Choi Department of Mathematics and Computer Science, Emory University • Define the task of “Character Identification”, a sub-task of entity linking. • Generate and analyze dialogue-specific corpus for the task. • Tackle the task via adaptation of existing coreference resolution systems. • Evaluate the performance of coreference resolution systems on the task. Objective • Given corpora, character identification can be solved with trained models • Existing coreference resolution systems can be modified to solve this task. • Models trained on one domain can be used to decode on other domains. • Models perform worse when trained with data of more than one domains. • Models trained on scene-level documents perform better in majority of the cases, though they failed at out-of-context character inferences. • Similar trends are observed after remapping the resultant coreferent chains from the systems in order to tackle character identification. Conclusion • Creation of a character identification specific entity linking system. • Resolution of the linking for collective and plural mentions. • Identification of disfluency markers and singleton mentions. • Generation of knowledge base from entity-centric attributes of mentions. Future Work Data Collection • Transcripts of TV shows, “Friends”(F) and “The Big Bang Theory”(BB), are collected from publically available websites and divided into scenes. Corpus Annotation • Corpus is double-annotated then adjudicated on Amazon Turk. Inter-Annotator Agreement • Quality of annotations is evaluated by their raw agreement and Cohen-Kappa scores. Corpus Creation Season Episode Scene Utterance Speaker Statement(s) Utterance text Figure 3. Structure of the corpus. Table 1. Statistics of the corpus. Epi/Sce/Spk: count of episodes, scenes, and speakers. UC/SC/WC: count of utterances, sentences, and words Figure 4. Template used on Amazon Turk for corpus annotation and adjudication Table 2. Inter-Annotator Agreement scores. F1p is a preliminary annotation trial done without context of the +-2 scenes and dynamic inferred speaker options • Character Identification is a sub-task of entity linking. It focuses on mapping mentions in context to one or more characters in a knowledge base. • The task is different from coreference resolution since the each coreferent chain does connect to an entity and is ambiguous for collective mentions. Task Definition Ross I told mom and dad last night, they seemed to take it pretty well. Monica Oh really, so that hysterical phone call I got from a woman at sobbing 3:00 A.M., "I'll never have grandchildren, I'll never have grandchildren." was what? A wrong number? MonicaJack JudyRoss Character Identification Figure 1. Task illustration of Character Identification. Ross I told mom and dad last night, they seemed to take it pretty well. Monica Oh really, so that hysterical phone call I got from a woman at sobbing 3:00 A.M., "I'll never have grandchildren, I'll never have grandchildren." was what? A wrong number? Coreference Resolution Figure 2. Task illustration of Coreference Resolution. • Linguistically-motivated rules, such as name entities, pronouns, and personal nouns, are used to select mentions from noun phrases. • Performance of the mention detection is manually evaluated by randomly examining 5% of the entire corpus. • The missed and error cases of the mention detection are further investigated. Mention Detection Table 3. Statistics of mentions found in our corpus. NE: Name entities. PRP: pronouns. PNN: personal nouns Table 4. Analysis on the performance of our mention detection. P: Precision. R: Recall. F: F-1 score. Table 1 Analogous phrases 2.06% 2 Misspelled pronouns 5.15% 5 Non-nominals 7.21% 7 Proper noun misses 9.28% 9 Interjection use of pronouns 14.43% 14 Common noun misses 14.43% 14 27% 27% 18% 14% 10% 4% Analogous phrases Misspelled pronouns Non-nominals Proper noun misses Interjection use of pronouns Common noun misses 1 Figure 5. Proportions of the misses and errors of the mention detection. Coreference Resolution • Stanford Multi-Sieve System is a rule-based system composed of multiple sieves of linguistic rules. It serves as the baseline of our task. • Stanford Entity-Centric System is a statistical system that extracts global entity-level features and constructs feature clusters for the resolution. • Only gold mentions are used for coreference resolutions. • Different combinations of the corpus are used to train the statistical system. Coreferent Chain Remapping • The resultant coreferent chains from the coreference resolution systems are remapped to an entity through voting based on the majority of mentions. • The referent of each mention is determined by pre-defined rules: 1. Character-related proper noun or named entity mention refers to the character of the proper noun or the named entity. 2. First-person pronoun or possessive pronoun mention refers to the character of the utterance contain the mention. 3. Collective pronoun or possessive pronoun refers to Collective group. Methodology Table 5. Corpus data split for training(TRN), developing(DEV) and testing(TST) • CoNLL’12 official scorer is used to evaluated immediate results of the coreference resolution systems with MUC, B3, and CEAFm metrics. • The remapped coreferent chains are evaluated in terms of purity scores. This represents the performance of the systems on character identification. Evaluation Table 6. Evaluations of the coreference resolution systems. Document episode/scene: each episode/scene is treated as a document. Table 7. Evaluations character identification after remapping the coreferent chains. FC/EC/UC: Found, expected, and unknown(%) clusters. UM: unknown(%) mentions.