SlideShare a Scribd company logo
Asking Clarifying Questions in
Open-Domain Information-
Seeking Conversations
Mohammad Aliannejadi(1), Hamed Zamani(2), Fabio Crestani(1), and W. Bruce Croft(2)
(1) Università della Svizzera italiana (USI), Switzerland
(2) University of Massachusetts Amherst, USA
© @dawnieando; @JeffD
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
5
6
7
8
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Can we ask questions to clarify
the user information needs?
Johannes Kiesel et al. Toward Voice Query Clarification. SIGIR 2018
Radlinski and Craswell. A Theoretical Framework for Conversational Search. CHIIR 2017
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
How to evaluate?
ClueWeb Collection
• A part of the Lemur Project
• A common web crawl (English) with 50M documents
• TREC Web Track 2009 – 2012
• Ad-hoc retrieval and diversification tasks
TREC facets
An offline evaluation methodology
• We assume that each user is interested in one facet per topic.
An offline evaluation methodology
• Let be the set of topics (queries).
• A collection of facet sets:
• includes all defined facets for topic .
• A collection of clarifying question sets:
• With including all clarifying questions relevant to topic .
• An offline evaluation requires defining .
•
Borrowed from the ClueWeb Collection
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Question Verification and Facet Linking
• Two main concerns:
• Precision: how is the quality of the collected clarifying questions?
• Recall: are all facets addressed by at least one clarifying question?
• Two expert annotators:
• Marked invalid and duplicate questions.
• Linked questions to the facets they found relevant.
• Facets with no questions: generated new questions relevant to them.
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
An offline evaluation methodology
• Let be the set of topics (queries).
• A collection of facet sets:
• includes all defined facets for topic .
• A collection of clarifying question sets:
• With including all clarifying questions relevant to topic .
• An offline evaluation requires defining .
•
Borrowed from the ClueWeb Collection
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Quality Check
• Regular quality checks on the collected answers.
• Manual checks on 10% of submissions per worker.
• If any invalid answer was observed, we then checked all the submissions of the
corresponding worker.
• Invalid answers were removed and workers banned from future tasks.
• Disabled copy/paste feature.
• Monitored keystrokes.
Qulac: Questions for Lack of Clarity
Qulac has two meanings in Persian:
• blizzard
• wonderful or masterpiece
© HBO
Learning to ask clarifying
questions
Question
Retrieval
Question Retrieval
• Task: Given a topic and a context (question-answer history), retrieve
clarifying questions.
• Desired objective: high recall
• Approaches:
• Term matching retrieval models: language models, BM25, RM3 (query
expansion)
• Learning to rank: LambdaMART, RankNet, neural ranking models (e.g., BERT)
Question Retrieval
Question
Retrieval
Question Selection
• Task: selecting a clarifying question that leads to retrieval improvement
• Objective: high precision (in retrieval)
• Approaches:
• Query performance prediction (QPP): predicting the retrieval performance after asking each
question (without answer) and selecting the one with the highest QPP.
• Learning to rank: defining a set of features for ranking questions. The features include QPP,
similarity to the topic, similarity to the context, etc.
• Neural ranking models: learning to rank with representation learning (e.g., BERT)
Question Selection
Asking only one good question improves the performance by over 100%.
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Asking Clarifying Questions in Open-Domain Information-Seeking Conversations
Case Study Negative answer; new information. Retrieval model fails.
Case Study Open question; new information.
Future Directions
• Utilizing positive and negative feedback for document retrieval.
• Joint modeling of question retrieval and selection.
• Question generation.
• Determining the number of questions to ask based on the system’s
confidence.
• Explore other ways of evaluating a system:
• Conversation turns;
• Retrieval performance.
Conclusions
• Asking clarifying questions in open-domain information-seeking conversations.
• Qulac: a collection for automatic offline evaluation of asking clarifying questions for
conversational IR.
• A simple yet effective retrieval framework.
• Asking only one good question improves the performance by over 100%!
• More improvement for:
• Shorter queries;
• Ambiguous queries.
Questions?
Qulac is publicly available at http://bit.ly/QulacData
Thanks to SIGIR for the student travel grant!

More Related Content

PDF
Systematic Literature Reviews and Systematic Mapping Studies
PDF
SelQA: A New Benchmark for Selection-based Question Answering
PPT
AI_Paper_Presentation
PDF
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
PPT
Shyam presentation prefinal
PDF
Presentation of Domain Specific Question Answering System Using N-gram Approach.
PDF
Detection of SOA Antipatterns
PDF
Qualitative Studies in Software Engineering - Interviews, Observation, Ground...
Systematic Literature Reviews and Systematic Mapping Studies
SelQA: A New Benchmark for Selection-based Question Answering
AI_Paper_Presentation
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
Shyam presentation prefinal
Presentation of Domain Specific Question Answering System Using N-gram Approach.
Detection of SOA Antipatterns
Qualitative Studies in Software Engineering - Interviews, Observation, Ground...

What's hot (20)

PPTX
Question Answering System using machine learning approach
PDF
FriendsQA: Open-domain Question Answering on TV Show Transcripts
PDF
2011 EASE - Motivation in Software Engineering: A Systematic Review Update
PPTX
Literature review
PDF
Empirical Evaluation of Active Learning in Recommender Systems
PDF
Using GradeMark to improve feedback and involve students in the marking process
PDF
User Personality and the New User Problem in a Context-Aware Point of Interes...
PPT
Using interactive models to enhance UML education
PPT
Learning-Based Evaluation of Visual Analytic Systems.
PDF
Ontology-Based Data Access Mapping Generation using Data, Schema, Query, and ...
PPTX
Introduction to Survey Data Quality
PPS
Multiple Response Questions - Allowing for chance in authentic assessments
PPTX
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in t...
PPTX
Towards Privacy-Preserving Evaluation for Information Retrieval Models over I...
PDF
Survey Research In Empirical Software Engineering
PDF
Apply chinese radicals into neural machine translation: deeper than character...
PPT
Proposal Templet 10page
PDF
AI Chatbot Service Framework based on Backpropagation Network for Predicting ...
PPT
Data Collection and Analysis Tools
PPTX
Developing high content image analysis software for biologists
Question Answering System using machine learning approach
FriendsQA: Open-domain Question Answering on TV Show Transcripts
2011 EASE - Motivation in Software Engineering: A Systematic Review Update
Literature review
Empirical Evaluation of Active Learning in Recommender Systems
Using GradeMark to improve feedback and involve students in the marking process
User Personality and the New User Problem in a Context-Aware Point of Interes...
Using interactive models to enhance UML education
Learning-Based Evaluation of Visual Analytic Systems.
Ontology-Based Data Access Mapping Generation using Data, Schema, Query, and ...
Introduction to Survey Data Quality
Multiple Response Questions - Allowing for chance in authentic assessments
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in t...
Towards Privacy-Preserving Evaluation for Information Retrieval Models over I...
Survey Research In Empirical Software Engineering
Apply chinese radicals into neural machine translation: deeper than character...
Proposal Templet 10page
AI Chatbot Service Framework based on Backpropagation Network for Predicting ...
Data Collection and Analysis Tools
Developing high content image analysis software for biologists
Ad

Similar to Asking Clarifying Questions in Open-Domain Information-Seeking Conversations (20)

PPTX
Query formulation process
PPTX
PDF
A SURVEY ON QUESTION AND ANSWER SYSTEM BY RETRIEVING THE DESCRIPTIONS USING L...
PPT
Is Search Broken?!
PPT
A data driven approach to query expansion in question answering
PDF
ACQSurvey (Poster)
PDF
Open domain Question Answering System - Research project in NLP
PPT
Where Does It Break?
PDF
Proposing a Semantic Multilingual Social Question Answering Service for Globa...
PDF
Clarification Questions Usefulness (Slides)
PDF
IRJET- Analysis of Question and Answering Recommendation System
DOC
Ibm piquant summary
PDF
Response quality-evaluation-in-heterogeneous-question-answering-system-a-blac...
PDF
A survey of automatic query expansion in information retrieval
PPTX
Machine Learned Relevance at A Large Scale Search Engine
PPTX
Techniques For Deep Query Understanding
PDF
Exploring session search
PDF
Application of hidden markov model in question answering systems
PPT
Tovek Presentation by Livio Costantini
PDF
Thesis presentation
Query formulation process
A SURVEY ON QUESTION AND ANSWER SYSTEM BY RETRIEVING THE DESCRIPTIONS USING L...
Is Search Broken?!
A data driven approach to query expansion in question answering
ACQSurvey (Poster)
Open domain Question Answering System - Research project in NLP
Where Does It Break?
Proposing a Semantic Multilingual Social Question Answering Service for Globa...
Clarification Questions Usefulness (Slides)
IRJET- Analysis of Question and Answering Recommendation System
Ibm piquant summary
Response quality-evaluation-in-heterogeneous-question-answering-system-a-blac...
A survey of automatic query expansion in information retrieval
Machine Learned Relevance at A Large Scale Search Engine
Techniques For Deep Query Understanding
Exploring session search
Application of hidden markov model in question answering systems
Tovek Presentation by Livio Costantini
Thesis presentation
Ad

Recently uploaded (20)

PPTX
The Effect of Human Resource Management Practice on Organizational Performanc...
PPTX
PHIL.-ASTRONOMY-AND-NAVIGATION of ..pptx
PPTX
Learning-Plan-5-Policies-and-Practices.pptx
PPTX
Human Mind & its character Characteristics
PPTX
Tablets And Capsule Preformulation Of Paracetamol
PPTX
water for all cao bang - a charity project
PDF
Swiggy’s Playbook: UX, Logistics & Monetization
PPTX
Intro to ISO 9001 2015.pptx wareness raising
PDF
Nykaa-Strategy-Case-Fixing-Retention-UX-and-D2C-Engagement (1).pdf
PPTX
Anesthesia and it's stage with mnemonic and images
PPTX
Introduction to Effective Communication.pptx
PPT
The Effect of Human Resource Management Practice on Organizational Performanc...
PPTX
BIOLOGY TISSUE PPT CLASS 9 PROJECT PUBLIC
PPTX
Emphasizing It's Not The End 08 06 2025.pptx
DOCX
ENGLISH PROJECT FOR BINOD BIHARI MAHTO KOYLANCHAL UNIVERSITY
DOC
学位双硕士UTAS毕业证,墨尔本理工学院毕业证留学硕士毕业证
PPTX
Hydrogel Based delivery Cancer Treatment
PPTX
Non-Verbal-Communication .mh.pdf_110245_compressed.pptx
PPTX
Project and change Managment: short video sequences for IBA
PPTX
worship songs, in any order, compilation
The Effect of Human Resource Management Practice on Organizational Performanc...
PHIL.-ASTRONOMY-AND-NAVIGATION of ..pptx
Learning-Plan-5-Policies-and-Practices.pptx
Human Mind & its character Characteristics
Tablets And Capsule Preformulation Of Paracetamol
water for all cao bang - a charity project
Swiggy’s Playbook: UX, Logistics & Monetization
Intro to ISO 9001 2015.pptx wareness raising
Nykaa-Strategy-Case-Fixing-Retention-UX-and-D2C-Engagement (1).pdf
Anesthesia and it's stage with mnemonic and images
Introduction to Effective Communication.pptx
The Effect of Human Resource Management Practice on Organizational Performanc...
BIOLOGY TISSUE PPT CLASS 9 PROJECT PUBLIC
Emphasizing It's Not The End 08 06 2025.pptx
ENGLISH PROJECT FOR BINOD BIHARI MAHTO KOYLANCHAL UNIVERSITY
学位双硕士UTAS毕业证,墨尔本理工学院毕业证留学硕士毕业证
Hydrogel Based delivery Cancer Treatment
Non-Verbal-Communication .mh.pdf_110245_compressed.pptx
Project and change Managment: short video sequences for IBA
worship songs, in any order, compilation

Asking Clarifying Questions in Open-Domain Information-Seeking Conversations

  • 1. Asking Clarifying Questions in Open-Domain Information- Seeking Conversations Mohammad Aliannejadi(1), Hamed Zamani(2), Fabio Crestani(1), and W. Bruce Croft(2) (1) Università della Svizzera italiana (USI), Switzerland (2) University of Massachusetts Amherst, USA
  • 5. 5
  • 6. 6
  • 7. 7
  • 8. 8
  • 12. Can we ask questions to clarify the user information needs? Johannes Kiesel et al. Toward Voice Query Clarification. SIGIR 2018 Radlinski and Craswell. A Theoretical Framework for Conversational Search. CHIIR 2017
  • 16. ClueWeb Collection • A part of the Lemur Project • A common web crawl (English) with 50M documents • TREC Web Track 2009 – 2012 • Ad-hoc retrieval and diversification tasks
  • 18. An offline evaluation methodology • We assume that each user is interested in one facet per topic.
  • 19. An offline evaluation methodology • Let be the set of topics (queries). • A collection of facet sets: • includes all defined facets for topic . • A collection of clarifying question sets: • With including all clarifying questions relevant to topic . • An offline evaluation requires defining . • Borrowed from the ClueWeb Collection
  • 23. Question Verification and Facet Linking • Two main concerns: • Precision: how is the quality of the collected clarifying questions? • Recall: are all facets addressed by at least one clarifying question? • Two expert annotators: • Marked invalid and duplicate questions. • Linked questions to the facets they found relevant. • Facets with no questions: generated new questions relevant to them.
  • 25. An offline evaluation methodology • Let be the set of topics (queries). • A collection of facet sets: • includes all defined facets for topic . • A collection of clarifying question sets: • With including all clarifying questions relevant to topic . • An offline evaluation requires defining . • Borrowed from the ClueWeb Collection
  • 27. Quality Check • Regular quality checks on the collected answers. • Manual checks on 10% of submissions per worker. • If any invalid answer was observed, we then checked all the submissions of the corresponding worker. • Invalid answers were removed and workers banned from future tasks. • Disabled copy/paste feature. • Monitored keystrokes.
  • 28. Qulac: Questions for Lack of Clarity Qulac has two meanings in Persian: • blizzard • wonderful or masterpiece © HBO
  • 29. Learning to ask clarifying questions
  • 31. Question Retrieval • Task: Given a topic and a context (question-answer history), retrieve clarifying questions. • Desired objective: high recall • Approaches: • Term matching retrieval models: language models, BM25, RM3 (query expansion) • Learning to rank: LambdaMART, RankNet, neural ranking models (e.g., BERT)
  • 34. Question Selection • Task: selecting a clarifying question that leads to retrieval improvement • Objective: high precision (in retrieval) • Approaches: • Query performance prediction (QPP): predicting the retrieval performance after asking each question (without answer) and selecting the one with the highest QPP. • Learning to rank: defining a set of features for ranking questions. The features include QPP, similarity to the topic, similarity to the context, etc. • Neural ranking models: learning to rank with representation learning (e.g., BERT)
  • 35. Question Selection Asking only one good question improves the performance by over 100%.
  • 40. Case Study Negative answer; new information. Retrieval model fails.
  • 41. Case Study Open question; new information.
  • 42. Future Directions • Utilizing positive and negative feedback for document retrieval. • Joint modeling of question retrieval and selection. • Question generation. • Determining the number of questions to ask based on the system’s confidence. • Explore other ways of evaluating a system: • Conversation turns; • Retrieval performance.
  • 43. Conclusions • Asking clarifying questions in open-domain information-seeking conversations. • Qulac: a collection for automatic offline evaluation of asking clarifying questions for conversational IR. • A simple yet effective retrieval framework. • Asking only one good question improves the performance by over 100%! • More improvement for: • Shorter queries; • Ambiguous queries.
  • 44. Questions? Qulac is publicly available at http://bit.ly/QulacData Thanks to SIGIR for the student travel grant!

Editor's Notes

  • #5: <number>
  • #13: make a frame of slides with logos... <number>
  • #14: make it more dialogue <number>
  • #15: put confidence here in the image <number>
  • #33: change this <number>
  • #43: remove one or two <number>
  • #44: more stress on confidence merge with future directions <number>