This paper proposes two approaches for document-level question answering: a pipeline approach and a confidence-based approach. The pipeline approach selects a single paragraph and extracts an answer from it. The confidence-based approach assigns confidence scores to answers from multiple paragraphs and returns the highest scoring answer. The paper experiments with different training methods for the confidence model and evaluates on several datasets, finding the shared normalization and no-answer option methods perform best. Error analysis shows the model still struggles with connecting statements across sentences and paragraphs.
Related topics: