SlideShare a Scribd company logo
End-to-end Plural Coreference
Resolution on TV Show Transcripts
Jose Coves
Advisor: Dr. Jinho D. Choi
Emory University, Department of Computer Science
Contents
■ Introduction
■ Coreference Resolution
■ End-to-end Coreference
■ Plural Coreference + metrics
■ Approaches
■ Results
Coreference Resolution
■ Coreference Resolution
– Find expressions that refer to the same entity
– Very important for higher-level NLP tasks
– Natural language understanding: QA, summarization, information extraction, etc.
– Unresolved fundamental NLP task
– Syntactical structures, speakers, sequential order, text comprehension
– Ambiguity: She told Monica she was smart. She told Joey she was smart.
– Entities: General (locations, objects, etc.) or specific (people).
– Mentions: Nested (The Wall of China). Plural (Mom and dad, they)
Corpus: Friends TV Show
■ Entities
– Known entities: Main characters. Eg. Joey is great.
– GENERIC: Characters whose identity is not revealed. Eg. I like the waitress.
– GENERAL: A class of people. Eg. The ideal girl doesn’t exist.
– OTHER: Identity unknown from local context. Eg. The guy next to me.
■ Annotation
– No nested plural entities (mom and dad).
– Plural mentions are not coreferent.
– Plural mentions added to coreferent entities’ clusters (they à mom, dad)
– GENERAL, OTHER à Singletons: One mention. Eg: I like women.
Neural Networks
■ FFNN – Feed output forward,
as input to next layer, without
forming a cycle
■ LSTM – Artificial RNN with
loops that allow information
to persist
■ CNN – Deep NN, extracts
most important features in
condensed form
End-to-end Coreference Resolution
■ Produce coreferent clusters by assigning antecedents to top span
■ No syntactic parser or mention-detector
■ Y(i) = {ε, 1, . . . , i−1}, for each top span i
■ Dummy Antecedent ε à Not a mention or not coreferent with any antecedent
■ Optimize marginal log-likelihood
■ Random initial prunning,
■ Only gold mentions get positive updates
End-to-end: Span representations
■ Max span width
■ Rank spans by mention score
■ Keep top K spans
End-to-end System
■ Element-wise similarity
■ Feature vector:
– speaker information
– distance factor
Higher-Order Coreference and Coarse-
to-Fine Inference
■ Higher-Order Coreference
– Expected antecedent
– Gate vector
– Update as weighted average
■ Coarse-to-Fine Inference
– Span ratio r, keep K = rT top spans, T is document length
– For each span, keep top C antecedents (Sc)
– Compute final coreference score S(i,j)
Plural Coreference Resolution
■ Uses gold mentions for test set predictions
■ Adds plural annotation to Friends corpus
■ Labeling for plural mentions
■ Clustering algorithm
■ Modify evaluation metrics
■ Feed all mention pairs into Agglomerative Convolutional Neural Network
ACNN classifier
Labeling and Clustering
■ For each span mj, look at each antecedent mi
■ L = S = Singular Antecedent: mi is singular. Assign mj to the cluster of mi
■ R = P = Plural Antecedent: mj is singular and mi is plural. Assign mi to mj’s cluster
Approach
■ End-to-end Neural Coreference + Plural Coreference Resolution
■ Adapt from CoNLL corpus to Friends corpus
– Nested mentions (nested mention detection, F1 72 vs 85)
– Singletons
– Plural speakers
– Character entities
– Plural mentions
■ Predict plural antecedents
■ Merge mentions into entity clusters
Plural speakers
■ Singular speakers à Same speaker binary flag vector
■ Plural speakers à Intersection of speakers, non-empty flag (speaker in common)
– Pros: captures existence of relationship
– Cons: no measure of match strength (exact vs partial)
■ Average speakers
– Assign embeddings to each speaker
– Average speakers’ embeddings
– Both mentions embeddings
– Pair-wise multiplication
Training Labels, Singletons
■ Baseline: End-to-end singular coreference resolution
– Modify gold labels for training set
– Pick ”head” mention, output singular clusters
– Evaluate on plural metrics
– Sort by appearance frequency
– Most popular, least popular, none. Eg. They à mom:2, dad:1
■ Singletons: Mentions not coreferent to other mentions
– Add left-over spans with mention score > threshold t (t=0)
– Only gold mentions receive + updates
Plural mentions
■ Singular Coreference
– Predict a coreferent antecedent for each mention
– Merge to the same cluster (transitive nature)
■ Plural Coreference
– Not transitive. Eg. {me, we}, {you, we} but not {me, you}
– Already have singular antecedents, need plural antecedents
– Predictions for all pairs of mentions (not just one per span)
Singularity
■ Singular + Plural antecedents à Revisit Higher-order coreference
– Weight Singular/Plural with Singularity S
■ Training loss = S*LossSingular + (1-S)*LossPlural 0.6 < S < 0.7
■ Antecedent labels
– Singular if span is singular (gold entity group size)
– Plural if span is plural and antecedent is singular
– Non-coreferent otherwise
Merging clusters + Many antecedents
■ Original clusters from baseline (singular antecedents): Add mi to cluster[mj]
– Then, for each plural prediction, add the span mj to cluster[mi]
– Example: I think we won. You did great. (we,you)àS, (I,we)àP. {you,we}, {I,we}
■ Only top singular antecedent for each span (or dummy)
– Softmax à Antecedents with score > dummy (no output label)
– Limit number of antecedents: maxplural and maxsingular (error accumulation)
– Example: I bought it for me, but we could share it.
Antecedent conflicts
■ Clustering: singular antecedents, then plural antecedents
■ Wrong predictions à Error propagation
■ New order for clustering
– Look at spans in order
– For each span, process its antecedents in order
■ Fix mention pair marked as singular and plural antecedent
– Pick the highest score
■ Example: I think we won. You did great.
■ Later spans, more antecedents
New Plural labels
■ Antecedent labels (Zhou and Choi)
– Singular if antecedent is singular
– Plural if antecedent is plural and span is singular
– Non-coreferent otherwise
– First singular, then plural antecedents
– New Plural
■ Reduce error propagation with span ordering
– New Plural + Ordering
Results
End-to-End Plural Coreference Resolution on TV Show Transcripts
■ F Base + plural New plural
Base + plural New plural
Analisys
■ Many Singular antecedents
– Pros: Helps with missed antecedents. Essential for plurals
Example: I, me, I. Missed (I, me). We, I, you. Need (We,I) and (I,you)
– Cons:
Example: I, we, you. Wrong (I,you)àS.
– f
Conclusion
■ Successfully adapt coreference model to Friends corpus
– Plural speakers, singletons, plural mentions
■ Modify singular coreference (end-to-end) for plural coreference
– Gradually identify weaknesses and improve performance
– Labeling techniques for plural antecedents
– Clustering of antecedents
■ First model to achieve end-to-end neural plural coreference resolution
Bibliography
Henry Y. Chen, Ethan Zhou, and Jinho D. Choi. Robust coreference resolution and entity linking on dialogues: Character identification on tv show transcripts. In Proceedings of
the 21st Conference on Computational Natural Language Learning (CoNLL 2017), pages 216–225. Association for Computational Linguistics, 2017. doi: 28 29
10.18653/v1/K17-1023. URL http://aclweb.org/anthology/K17-1023.
Yu-Hsin Chen and Jinho D. Choi. Character identification on multiparty conversation: Identifying mentions of characters in tv shows. In Proceedings of the 17th Annual
Meeting of the Special Interest Group on Discourse and Dialogue, pages 90–100. Association for Computational Linguistics, 2016. doi: 10.18653/v1/W16-3612. URL
http://aclweb.org/anthology/W16-3612.
Kevin Clark and Christopher D. Manning. Deep reinforcement learning for mention-ranking coreference models. In Proceedings of the 2016 Conference on Empirical Methods
in Natural Language Processing, pages 2256–2262. Association for Computational Linguistics, 2016. doi: 10.18653/v1/D16-1245. URL http: //aclweb.org/anthology/D16-1245.
Kevin Clark and Christopher D. Manning. Improving coreference resolution by learning entity-level distributed representations. CoRR, abs/1606.01323, 2016. URL
http://arxiv.org/abs/1606.01323.
Arzoo Katiyar and Claire Cardie. Nested named entity recognition revisited. In Proceedings of the 2018 Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 861–871, New Orleans, Louisiana, June 2018. Association for 30 Computational
Linguistics. doi: 10.18653/v1/N18-1079. URL http://www. aclweb.org/anthology/N18-1079.
Kenton Lee, Luheng He, Mike Lewis, and Luke Zettlemoyer. End-to-end neural coreference resolution. CoRR, abs/1707.07045, 2017. URL http://arxiv.org/ abs/1707.07045.
Kenton Lee, Luheng He, and Luke Zettlemoyer. Higher-order coreference resolution with coarse-to-fine inference. CoRR, abs/1804.05392, 2018. URL http://arxiv.
org/abs/1804.05392.
Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Olga Uryupina, and Yuchen Zhang. Conll-2012 shared task: Modeling multilingual unrestricted coreference in ontonotes.
In Joint Conference on EMNLP and CoNLL - Shared Task, CoNLL ’12, pages 1–40, Stroudsburg, PA, USA, 2012. Association for Computational Linguistics. URL
http://dl.acm.org/citation.cfm?id=2391181.2391183.
Sam Wiseman, Alexander M. Rush, and Stuart M. Shieber. Learning global features for coreference resolution. CoRR, abs/1604.03035, 2016. URL http://arxiv.
org/abs/1604.03035
Ethan Zhou and Jinho D. Choi. They exist! introducing plural mentions to coreference resolution and entity linking. In Proceedings of the 27th International Conference on 31
Computational Linguistics, pages 24–34. Association for Computational Linguistics, 2018. URL http://aclweb.org/anthology/C18-1003.
https://nlp.stanford.edu/projects/coref.shtml
Metrics

More Related Content

PPTX
Coreference_Resolution in Natural language processing
PDF
Real-time Coreference Resolution for Dialogue Understanding
PDF
Character Identification on Multi-party Dialogues
PDF
Improving Machine Learning Approaches to Coreference Resolution
DOC
Coreference Resolution using Hybrid Approach
PDF
Event Coreference Resolution using Mincut based Graph Clustering
PDF
Anaphora Resolution
PDF
Robust Coreference Resolution and Entity Linking on Dialogues: Character Iden...
Coreference_Resolution in Natural language processing
Real-time Coreference Resolution for Dialogue Understanding
Character Identification on Multi-party Dialogues
Improving Machine Learning Approaches to Coreference Resolution
Coreference Resolution using Hybrid Approach
Event Coreference Resolution using Mincut based Graph Clustering
Anaphora Resolution
Robust Coreference Resolution and Entity Linking on Dialogues: Character Iden...

Similar to End-to-End Plural Coreference Resolution on TV Show Transcripts (20)

PDF
Character Identification on Multiparty Conversation: Identifying Mentions of ...
PDF
A Pilot Study On Computer-Aided Coreference Annotation
PPT
Coreference Resolution
PPT
anaphora resolution natural language processing.ppt
PDF
Monolingual Phrase Alignment on Parse Forests (EMNLP2017 presentation)
PDF
An Entity-Driven Recursive Neural Network Model for Chinese Discourse Coheren...
PPTX
anaphora resolution.pptx
PDF
Tutorial on Coreference Resolution
ODP
Reference Scope Identification in Citing Sentences
PDF
A^2_Poster
PDF
PPTX
SANAPHOR: Ontology-based Coreference Resolution
PDF
A Graph-based Cross-lingual Projection Approach for Weakly Supervised Relatio...
PDF
Domain Specific Named Entity Recognition Using Supervised Approach
PPTX
Predicting sentential semantic compatibility for aggregation in text to-text ...
PPT
lect26-em.ppt
PPTX
Coreference Extraction from Identric’s Documents - Solution of Datathon 2018
PDF
The Myth of Higher-Order Inference in Coreference Resolution
PDF
Neural Network in Knowledge Bases
PDF
Syntax Directed Translation PPTs for Third Year CSE
Character Identification on Multiparty Conversation: Identifying Mentions of ...
A Pilot Study On Computer-Aided Coreference Annotation
Coreference Resolution
anaphora resolution natural language processing.ppt
Monolingual Phrase Alignment on Parse Forests (EMNLP2017 presentation)
An Entity-Driven Recursive Neural Network Model for Chinese Discourse Coheren...
anaphora resolution.pptx
Tutorial on Coreference Resolution
Reference Scope Identification in Citing Sentences
A^2_Poster
SANAPHOR: Ontology-based Coreference Resolution
A Graph-based Cross-lingual Projection Approach for Weakly Supervised Relatio...
Domain Specific Named Entity Recognition Using Supervised Approach
Predicting sentential semantic compatibility for aggregation in text to-text ...
lect26-em.ppt
Coreference Extraction from Identric’s Documents - Solution of Datathon 2018
The Myth of Higher-Order Inference in Coreference Resolution
Neural Network in Knowledge Bases
Syntax Directed Translation PPTs for Third Year CSE
Ad

More from Jinho Choi (20)

PDF
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
PDF
Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP D...
PDF
Competence-Level Prediction and Resume & Job Description Matching Using Conte...
PDF
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
PDF
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
PDF
Abstract Meaning Representation
PDF
Semantic Role Labeling
PDF
CKY Parsing
PDF
CS329 - WordNet Similarities
PDF
CS329 - Lexical Relations
PDF
Automatic Knowledge Base Expansion for Dialogue Management
PDF
Attention is All You Need for AMR Parsing
PDF
Graph-to-Text Generation and its Applications to Dialogue
PDF
Topological Sort
PDF
Tries - Put
PDF
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
PDF
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
PDF
How to make Emora talk about Sports Intelligently
PDF
Text-to-SQL with Data-Driven Templates
PDF
Resume Classification with Term Attention Embeddings
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP D...
Competence-Level Prediction and Resume & Job Description Matching Using Conte...
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
Abstract Meaning Representation
Semantic Role Labeling
CKY Parsing
CS329 - WordNet Similarities
CS329 - Lexical Relations
Automatic Knowledge Base Expansion for Dialogue Management
Attention is All You Need for AMR Parsing
Graph-to-Text Generation and its Applications to Dialogue
Topological Sort
Tries - Put
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
How to make Emora talk about Sports Intelligently
Text-to-SQL with Data-Driven Templates
Resume Classification with Term Attention Embeddings
Ad

Recently uploaded (20)

PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PDF
Machine learning based COVID-19 study performance prediction
PDF
Encapsulation theory and applications.pdf
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
A comparative analysis of optical character recognition models for extracting...
PPTX
Big Data Technologies - Introduction.pptx
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Assigned Numbers - 2025 - Bluetooth® Document
PDF
Empathic Computing: Creating Shared Understanding
PPTX
1. Introduction to Computer Programming.pptx
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
Spectral efficient network and resource selection model in 5G networks
PPT
Teaching material agriculture food technology
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Encapsulation_ Review paper, used for researhc scholars
Accuracy of neural networks in brain wave diagnosis of schizophrenia
Machine learning based COVID-19 study performance prediction
Encapsulation theory and applications.pdf
Digital-Transformation-Roadmap-for-Companies.pptx
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Reach Out and Touch Someone: Haptics and Empathic Computing
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
MIND Revenue Release Quarter 2 2025 Press Release
A comparative analysis of optical character recognition models for extracting...
Big Data Technologies - Introduction.pptx
Network Security Unit 5.pdf for BCA BBA.
Assigned Numbers - 2025 - Bluetooth® Document
Empathic Computing: Creating Shared Understanding
1. Introduction to Computer Programming.pptx
The Rise and Fall of 3GPP – Time for a Sabbatical?
Spectral efficient network and resource selection model in 5G networks
Teaching material agriculture food technology
Per capita expenditure prediction using model stacking based on satellite ima...
Encapsulation_ Review paper, used for researhc scholars

End-to-End Plural Coreference Resolution on TV Show Transcripts

  • 1. End-to-end Plural Coreference Resolution on TV Show Transcripts Jose Coves Advisor: Dr. Jinho D. Choi Emory University, Department of Computer Science
  • 2. Contents ■ Introduction ■ Coreference Resolution ■ End-to-end Coreference ■ Plural Coreference + metrics ■ Approaches ■ Results
  • 3. Coreference Resolution ■ Coreference Resolution – Find expressions that refer to the same entity – Very important for higher-level NLP tasks – Natural language understanding: QA, summarization, information extraction, etc. – Unresolved fundamental NLP task – Syntactical structures, speakers, sequential order, text comprehension – Ambiguity: She told Monica she was smart. She told Joey she was smart. – Entities: General (locations, objects, etc.) or specific (people). – Mentions: Nested (The Wall of China). Plural (Mom and dad, they)
  • 4. Corpus: Friends TV Show ■ Entities – Known entities: Main characters. Eg. Joey is great. – GENERIC: Characters whose identity is not revealed. Eg. I like the waitress. – GENERAL: A class of people. Eg. The ideal girl doesn’t exist. – OTHER: Identity unknown from local context. Eg. The guy next to me. ■ Annotation – No nested plural entities (mom and dad). – Plural mentions are not coreferent. – Plural mentions added to coreferent entities’ clusters (they à mom, dad) – GENERAL, OTHER à Singletons: One mention. Eg: I like women.
  • 5. Neural Networks ■ FFNN – Feed output forward, as input to next layer, without forming a cycle ■ LSTM – Artificial RNN with loops that allow information to persist ■ CNN – Deep NN, extracts most important features in condensed form
  • 6. End-to-end Coreference Resolution ■ Produce coreferent clusters by assigning antecedents to top span ■ No syntactic parser or mention-detector ■ Y(i) = {ε, 1, . . . , i−1}, for each top span i ■ Dummy Antecedent ε à Not a mention or not coreferent with any antecedent ■ Optimize marginal log-likelihood ■ Random initial prunning, ■ Only gold mentions get positive updates
  • 7. End-to-end: Span representations ■ Max span width ■ Rank spans by mention score ■ Keep top K spans
  • 8. End-to-end System ■ Element-wise similarity ■ Feature vector: – speaker information – distance factor
  • 9. Higher-Order Coreference and Coarse- to-Fine Inference ■ Higher-Order Coreference – Expected antecedent – Gate vector – Update as weighted average ■ Coarse-to-Fine Inference – Span ratio r, keep K = rT top spans, T is document length – For each span, keep top C antecedents (Sc) – Compute final coreference score S(i,j)
  • 10. Plural Coreference Resolution ■ Uses gold mentions for test set predictions ■ Adds plural annotation to Friends corpus ■ Labeling for plural mentions ■ Clustering algorithm ■ Modify evaluation metrics ■ Feed all mention pairs into Agglomerative Convolutional Neural Network
  • 12. Labeling and Clustering ■ For each span mj, look at each antecedent mi ■ L = S = Singular Antecedent: mi is singular. Assign mj to the cluster of mi ■ R = P = Plural Antecedent: mj is singular and mi is plural. Assign mi to mj’s cluster
  • 13. Approach ■ End-to-end Neural Coreference + Plural Coreference Resolution ■ Adapt from CoNLL corpus to Friends corpus – Nested mentions (nested mention detection, F1 72 vs 85) – Singletons – Plural speakers – Character entities – Plural mentions ■ Predict plural antecedents ■ Merge mentions into entity clusters
  • 14. Plural speakers ■ Singular speakers à Same speaker binary flag vector ■ Plural speakers à Intersection of speakers, non-empty flag (speaker in common) – Pros: captures existence of relationship – Cons: no measure of match strength (exact vs partial) ■ Average speakers – Assign embeddings to each speaker – Average speakers’ embeddings – Both mentions embeddings – Pair-wise multiplication
  • 15. Training Labels, Singletons ■ Baseline: End-to-end singular coreference resolution – Modify gold labels for training set – Pick ”head” mention, output singular clusters – Evaluate on plural metrics – Sort by appearance frequency – Most popular, least popular, none. Eg. They à mom:2, dad:1 ■ Singletons: Mentions not coreferent to other mentions – Add left-over spans with mention score > threshold t (t=0) – Only gold mentions receive + updates
  • 16. Plural mentions ■ Singular Coreference – Predict a coreferent antecedent for each mention – Merge to the same cluster (transitive nature) ■ Plural Coreference – Not transitive. Eg. {me, we}, {you, we} but not {me, you} – Already have singular antecedents, need plural antecedents – Predictions for all pairs of mentions (not just one per span)
  • 17. Singularity ■ Singular + Plural antecedents à Revisit Higher-order coreference – Weight Singular/Plural with Singularity S ■ Training loss = S*LossSingular + (1-S)*LossPlural 0.6 < S < 0.7 ■ Antecedent labels – Singular if span is singular (gold entity group size) – Plural if span is plural and antecedent is singular – Non-coreferent otherwise
  • 18. Merging clusters + Many antecedents ■ Original clusters from baseline (singular antecedents): Add mi to cluster[mj] – Then, for each plural prediction, add the span mj to cluster[mi] – Example: I think we won. You did great. (we,you)àS, (I,we)àP. {you,we}, {I,we} ■ Only top singular antecedent for each span (or dummy) – Softmax à Antecedents with score > dummy (no output label) – Limit number of antecedents: maxplural and maxsingular (error accumulation) – Example: I bought it for me, but we could share it.
  • 19. Antecedent conflicts ■ Clustering: singular antecedents, then plural antecedents ■ Wrong predictions à Error propagation ■ New order for clustering – Look at spans in order – For each span, process its antecedents in order ■ Fix mention pair marked as singular and plural antecedent – Pick the highest score ■ Example: I think we won. You did great. ■ Later spans, more antecedents
  • 20. New Plural labels ■ Antecedent labels (Zhou and Choi) – Singular if antecedent is singular – Plural if antecedent is plural and span is singular – Non-coreferent otherwise – First singular, then plural antecedents – New Plural ■ Reduce error propagation with span ordering – New Plural + Ordering
  • 23. ■ F Base + plural New plural
  • 24. Base + plural New plural
  • 25. Analisys ■ Many Singular antecedents – Pros: Helps with missed antecedents. Essential for plurals Example: I, me, I. Missed (I, me). We, I, you. Need (We,I) and (I,you) – Cons: Example: I, we, you. Wrong (I,you)àS. – f
  • 26. Conclusion ■ Successfully adapt coreference model to Friends corpus – Plural speakers, singletons, plural mentions ■ Modify singular coreference (end-to-end) for plural coreference – Gradually identify weaknesses and improve performance – Labeling techniques for plural antecedents – Clustering of antecedents ■ First model to achieve end-to-end neural plural coreference resolution
  • 27. Bibliography Henry Y. Chen, Ethan Zhou, and Jinho D. Choi. Robust coreference resolution and entity linking on dialogues: Character identification on tv show transcripts. In Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017), pages 216–225. Association for Computational Linguistics, 2017. doi: 28 29 10.18653/v1/K17-1023. URL http://aclweb.org/anthology/K17-1023. Yu-Hsin Chen and Jinho D. Choi. Character identification on multiparty conversation: Identifying mentions of characters in tv shows. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 90–100. Association for Computational Linguistics, 2016. doi: 10.18653/v1/W16-3612. URL http://aclweb.org/anthology/W16-3612. Kevin Clark and Christopher D. Manning. Deep reinforcement learning for mention-ranking coreference models. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2256–2262. Association for Computational Linguistics, 2016. doi: 10.18653/v1/D16-1245. URL http: //aclweb.org/anthology/D16-1245. Kevin Clark and Christopher D. Manning. Improving coreference resolution by learning entity-level distributed representations. CoRR, abs/1606.01323, 2016. URL http://arxiv.org/abs/1606.01323. Arzoo Katiyar and Claire Cardie. Nested named entity recognition revisited. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 861–871, New Orleans, Louisiana, June 2018. Association for 30 Computational Linguistics. doi: 10.18653/v1/N18-1079. URL http://www. aclweb.org/anthology/N18-1079. Kenton Lee, Luheng He, Mike Lewis, and Luke Zettlemoyer. End-to-end neural coreference resolution. CoRR, abs/1707.07045, 2017. URL http://arxiv.org/ abs/1707.07045. Kenton Lee, Luheng He, and Luke Zettlemoyer. Higher-order coreference resolution with coarse-to-fine inference. CoRR, abs/1804.05392, 2018. URL http://arxiv. org/abs/1804.05392. Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Olga Uryupina, and Yuchen Zhang. Conll-2012 shared task: Modeling multilingual unrestricted coreference in ontonotes. In Joint Conference on EMNLP and CoNLL - Shared Task, CoNLL ’12, pages 1–40, Stroudsburg, PA, USA, 2012. Association for Computational Linguistics. URL http://dl.acm.org/citation.cfm?id=2391181.2391183. Sam Wiseman, Alexander M. Rush, and Stuart M. Shieber. Learning global features for coreference resolution. CoRR, abs/1604.03035, 2016. URL http://arxiv. org/abs/1604.03035 Ethan Zhou and Jinho D. Choi. They exist! introducing plural mentions to coreference resolution and entity linking. In Proceedings of the 27th International Conference on 31 Computational Linguistics, pages 24–34. Association for Computational Linguistics, 2018. URL http://aclweb.org/anthology/C18-1003. https://nlp.stanford.edu/projects/coref.shtml