SlideShare a Scribd company logo
Learning-Based Evaluation of Visual
Analytics Systems


   Remco Chang, Caroline Ziemkiewicz, Roman
     Pyzh, Joseph Kielman*, William Ribarsky

    UNC Charlotte                    *Department of Homeland
    Charlotte Visualization Center   Security
Why Another Evaluation Method?

• Based on a discussion with Joe Kielman (DHS)
   – Why is it difficult for agencies like the DHS to adopt
     and use visual analytics systems?


• Most existing metrics are not indicative of
  success of adoption
   –   Task completion time
   –   Errors
   –   Subjective preferences
   –   Etc.
Current Methods

• Methods for evaluating visual analytics
  systems have been proposed. Each has its
  unique perspective and goal. For example:

  – Insight-based Evaluation (North et al.)
  – Productivity-based Evaluation (Scholtz)
  – MILC -- Multi-dimensional in-depth long-term case
    studies (Schneiderman, Plaisant)
  – Grounded Evaluation (Isenberg et al.)
Our Goal for Evaluation

• What Joe wants is:
   – Proof that the user of the visual analytics system can gain
     proficiency in solving a problem using the system

   – By using the VA system, show that a user can gradually
     change from being a “novice” to becoming an “expert”

• In other words, Joe wants proof that by using the VA
  system, the user is gaining knowledge…
   – The goal of visualization is to gain insight and knowledge
     (ViSC report, 1987) (Illuminating the Path)
Learning-Based Evaluation

• In light of this goal, we propose a “learning-based
  evaluation” that attempts to directly test the
  amount of knowledge gained by its user.

• The idea is try to determine how much the user
  has learned after spending time using a VA
  system by:
   – Giving a user a similar but different task.
   – Directly testing if the user has gained proficiency in
     the subject matter.
Current Method
Our Proposed Method
Types of Learning

• In designing either a new task or the
  questionnaire, it is important to differentiate
  and isolate what is being tested:

  – Knowledge gained about the Interface
  – Knowledge gained about the data
  – Knowledge gained about the task (domain)
iPCA Example
• iPCA stands for
  “interactive Principle
  Component Analysis”. By
  using it, the user can learn
  about:
   – The interface
   – The dataset
      • relationships within the
        data
   – The task
      • What is principle
        component analysis, and
      • How can I use principle
        component analysis to solve
        other problems?
Application to the VAST Challenge

• Current method:
  – Give participants a dataset and a problem
  – Ask participants to develop VA systems to solve
    the problem
  – Ask participants to describe their systems and
    analytical methods
  – Judges score each submission based on the
    developed systems and their applicability to the
    problem
Application to the VAST Challenge

• Proposed method:
  – Give participants a dataset and a problem
  – Ask participants to develop VA system to solve the
    problem
  – Ask participants to bring their systems to VisWeek
  – Give participants a similar, but different dataset and
    problem
  – Ask participants to solve the new problem using
    their VA systems
  – Judges score each participant based on the
    effectiveness of each system in solving the new task.
Types of Learning

• In designing either a new task or the
  questionnaire, it is important to differentiate
  and isolate what is being tested:

  – Knowledge gained about the Interface
  – Knowledge gained about the data
  – Knowledge gained about the task (domain)
Discussion/Conclusion
• This learning-based method seems simple and obvious
  because it really is. Teachers have been doing this for
  ages.

• The method is not unique. There are many aspects of
  this proposed method that are similar to existing
  methods. In spirit, we are all looking to address the
  same problem.

• The difference is the perspective. If we think about the
  problem from the perspective of a client (e.g., Joe at
  DHS), what they look for in evaluation results currently
  are not the same as what we as researchers give them.
Future Work

• Integrate the proposed learning-based
  method to:
  – Grounded Evaluation
  – Long term effects (MILC)
Thank you!




         rchang@uncc.edu
http://www.viscenter.uncc.edu/~rchang
The Classroom Analogy

• Say you’re a math teacher in middle school, and
  you’re trying to decide which text book to use,
  the blue one or the red one. You can:
  – Ask your friends which book is better
     • Analogous to an “expert-based evaluation”. Problem is that
       the sample size is typically small, and the results difficult to
       replicate.
  – Ask your students which book they like
     • Analogous to subjective preferences. The issue here is that
       the students can prefer the blue text book because its blue.
  – Test which text book is more effective by giving the
    students tests.

More Related Content

PDF
ch5 Issues based metrics
PPTX
4 reasons why you need online assessments
PDF
Examview Features and benefits
PPT
Data Collection and Analysis Tools
PPSX
Caveon webinar series Standard Setting for the 21st Century, Using Informa...
PDF
Best Practices in Recommender System Challenges
PDF
Flow chart redesign
PPT
Getting Them There: A Small-Scale Usability Test of a University Library Website
ch5 Issues based metrics
4 reasons why you need online assessments
Examview Features and benefits
Data Collection and Analysis Tools
Caveon webinar series Standard Setting for the 21st Century, Using Informa...
Best Practices in Recommender System Challenges
Flow chart redesign
Getting Them There: A Small-Scale Usability Test of a University Library Website

What's hot (20)

PPTX
Classsourcing: Crowd-Based Validation of Question-Answer Learning Objects @ I...
PPT
Usability Evaluation in Educational Technology
PPTX
Consumer Oriented Evaluation Ppt
PPTX
Nonnegative matrix-fact
PPS
Multiple Response Questions - Allowing for chance in authentic assessments
PPTX
pilot testing of questionnaire
PDF
Active Learning in Collaborative Filtering Recommender Systems : a Survey
PPS
Topic 7 Product Evaluation
PPTX
Introduction to Survey Data Quality
PPTX
Overseas mkt research 2
PDF
Design of Experiments (DoE) Seminar Outline
PDF
Meta-study results of participatory processes
PDF
Using GradeMark to improve feedback and involve students in the marking process
PPTX
Usability Testing
PDF
Beat the odds evaluation model table
PPTX
Analyzing data the data is coming
PPTX
Prediction-Next-Term Student Performance Prediction: A Recommender Systems Ap...
PPT
FYP ppt
PPTX
Opening up multiple choice - assessing with confidence
PPTX
Ch. 7 finish and review
Classsourcing: Crowd-Based Validation of Question-Answer Learning Objects @ I...
Usability Evaluation in Educational Technology
Consumer Oriented Evaluation Ppt
Nonnegative matrix-fact
Multiple Response Questions - Allowing for chance in authentic assessments
pilot testing of questionnaire
Active Learning in Collaborative Filtering Recommender Systems : a Survey
Topic 7 Product Evaluation
Introduction to Survey Data Quality
Overseas mkt research 2
Design of Experiments (DoE) Seminar Outline
Meta-study results of participatory processes
Using GradeMark to improve feedback and involve students in the marking process
Usability Testing
Beat the odds evaluation model table
Analyzing data the data is coming
Prediction-Next-Term Student Performance Prediction: A Recommender Systems Ap...
FYP ppt
Opening up multiple choice - assessing with confidence
Ch. 7 finish and review
Ad

Viewers also liked (8)

PDF
How is a graphic like pumpkin pie? A framework for analysis and critique of v...
PPT
Many Roads Lead to Rome. Mapping Users’ Problem Solving Strategies.
PPTX
Proposed Working Memory Measures for Evaluating Information Visualization Tools.
PDF
Look Before You Link: Eye Tracking in Multiple Coordinated View Visualization.
PPTX
Implications of Individual Differences on Evaluating Information Visualizatio...
PPT
Pragmatic Challenges in the Evaluation of Interactive Visualization Systems.
PPT
Evaluating Information Visualization in Large Companies: Challenges, Experien...
PPTX
Comparative Evaluation of Two Interface Tools in Performing Visual Analytics ...
How is a graphic like pumpkin pie? A framework for analysis and critique of v...
Many Roads Lead to Rome. Mapping Users’ Problem Solving Strategies.
Proposed Working Memory Measures for Evaluating Information Visualization Tools.
Look Before You Link: Eye Tracking in Multiple Coordinated View Visualization.
Implications of Individual Differences on Evaluating Information Visualizatio...
Pragmatic Challenges in the Evaluation of Interactive Visualization Systems.
Evaluating Information Visualization in Large Companies: Challenges, Experien...
Comparative Evaluation of Two Interface Tools in Performing Visual Analytics ...
Ad

Similar to Learning-Based Evaluation of Visual Analytic Systems. (20)

PDF
NEAFCS 2012 Financial Ed eval-assessment-impact-o'neill-09-12
PDF
Empowering youth to be evaluators: Involving Young People in Evaluating Infor...
PDF
Field Studies as evaluation method for socio-technical interventions in Techn...
PPTX
Mmig talk jan 245 2011
PPTX
The perils of measurement
PPTX
Researchsaysls2017
PDF
AFCPE 2016 Symposium workshop-Measuring & Reporting Impact of Financial Educa...
PDF
(eBook PDF) Handbook of Practical Program Evaluation 4th Edition
PPTX
Keynote Designing Assessment, Assessing Instructional Design
PDF
Contribution300 a
PPTX
What will they need? Pre-assessment techniques for instruction session.
PDF
ACM ITICSE 2014 - Talk on Motivational Active Learning
PDF
Learning theories
PDF
Play’n’Learn: A Continuous KM Improvement Approach using FSM methods
PDF
Developmental Evaluation and the Graduate Student Researcher
PDF
PPT
2008 regional educational laboratory board of directors (rel midwest)
DOCX
U4D2 - Action Learning Methods and Assessment Planning ToolsIm.docx
PPTX
Seeking Evidence of Impact: Answering "How Do We Know?"
PPT
Assessments for Programs and Learning
NEAFCS 2012 Financial Ed eval-assessment-impact-o'neill-09-12
Empowering youth to be evaluators: Involving Young People in Evaluating Infor...
Field Studies as evaluation method for socio-technical interventions in Techn...
Mmig talk jan 245 2011
The perils of measurement
Researchsaysls2017
AFCPE 2016 Symposium workshop-Measuring & Reporting Impact of Financial Educa...
(eBook PDF) Handbook of Practical Program Evaluation 4th Edition
Keynote Designing Assessment, Assessing Instructional Design
Contribution300 a
What will they need? Pre-assessment techniques for instruction session.
ACM ITICSE 2014 - Talk on Motivational Active Learning
Learning theories
Play’n’Learn: A Continuous KM Improvement Approach using FSM methods
Developmental Evaluation and the Graduate Student Researcher
2008 regional educational laboratory board of directors (rel midwest)
U4D2 - Action Learning Methods and Assessment Planning ToolsIm.docx
Seeking Evidence of Impact: Answering "How Do We Know?"
Assessments for Programs and Learning

More from BELIV Workshop (9)

PDF
Towards Information-Theoretic Visualization Evaluation Measure: A Practical e...
PPT
Is Your User Hunting or Gathering Insights? Identifying Insight Drivers Acros...
KEY
A Descriptive Model of Visual Scanning.
PPT
Generating a synthetic video dataset
PPTX
Beyond system logging: human logging for evaluating information visualization.
PPT
Scanning Between Graph Visualizations: An Eye Tracking Evaluation.
PPTX
Focus Groups for Functional InfoVis Prototype Evaluation: A Case Study.
PDF
Visualization Evaluation of the Masses, by the Masses, and for the Masses.
PPTX
BELIV'10 Keynote: Conceptual and Practical Challenges in InfoViz Evaluations
Towards Information-Theoretic Visualization Evaluation Measure: A Practical e...
Is Your User Hunting or Gathering Insights? Identifying Insight Drivers Acros...
A Descriptive Model of Visual Scanning.
Generating a synthetic video dataset
Beyond system logging: human logging for evaluating information visualization.
Scanning Between Graph Visualizations: An Eye Tracking Evaluation.
Focus Groups for Functional InfoVis Prototype Evaluation: A Case Study.
Visualization Evaluation of the Masses, by the Masses, and for the Masses.
BELIV'10 Keynote: Conceptual and Practical Challenges in InfoViz Evaluations

Learning-Based Evaluation of Visual Analytic Systems.

  • 1. Learning-Based Evaluation of Visual Analytics Systems Remco Chang, Caroline Ziemkiewicz, Roman Pyzh, Joseph Kielman*, William Ribarsky UNC Charlotte *Department of Homeland Charlotte Visualization Center Security
  • 2. Why Another Evaluation Method? • Based on a discussion with Joe Kielman (DHS) – Why is it difficult for agencies like the DHS to adopt and use visual analytics systems? • Most existing metrics are not indicative of success of adoption – Task completion time – Errors – Subjective preferences – Etc.
  • 3. Current Methods • Methods for evaluating visual analytics systems have been proposed. Each has its unique perspective and goal. For example: – Insight-based Evaluation (North et al.) – Productivity-based Evaluation (Scholtz) – MILC -- Multi-dimensional in-depth long-term case studies (Schneiderman, Plaisant) – Grounded Evaluation (Isenberg et al.)
  • 4. Our Goal for Evaluation • What Joe wants is: – Proof that the user of the visual analytics system can gain proficiency in solving a problem using the system – By using the VA system, show that a user can gradually change from being a “novice” to becoming an “expert” • In other words, Joe wants proof that by using the VA system, the user is gaining knowledge… – The goal of visualization is to gain insight and knowledge (ViSC report, 1987) (Illuminating the Path)
  • 5. Learning-Based Evaluation • In light of this goal, we propose a “learning-based evaluation” that attempts to directly test the amount of knowledge gained by its user. • The idea is try to determine how much the user has learned after spending time using a VA system by: – Giving a user a similar but different task. – Directly testing if the user has gained proficiency in the subject matter.
  • 8. Types of Learning • In designing either a new task or the questionnaire, it is important to differentiate and isolate what is being tested: – Knowledge gained about the Interface – Knowledge gained about the data – Knowledge gained about the task (domain)
  • 9. iPCA Example • iPCA stands for “interactive Principle Component Analysis”. By using it, the user can learn about: – The interface – The dataset • relationships within the data – The task • What is principle component analysis, and • How can I use principle component analysis to solve other problems?
  • 10. Application to the VAST Challenge • Current method: – Give participants a dataset and a problem – Ask participants to develop VA systems to solve the problem – Ask participants to describe their systems and analytical methods – Judges score each submission based on the developed systems and their applicability to the problem
  • 11. Application to the VAST Challenge • Proposed method: – Give participants a dataset and a problem – Ask participants to develop VA system to solve the problem – Ask participants to bring their systems to VisWeek – Give participants a similar, but different dataset and problem – Ask participants to solve the new problem using their VA systems – Judges score each participant based on the effectiveness of each system in solving the new task.
  • 12. Types of Learning • In designing either a new task or the questionnaire, it is important to differentiate and isolate what is being tested: – Knowledge gained about the Interface – Knowledge gained about the data – Knowledge gained about the task (domain)
  • 13. Discussion/Conclusion • This learning-based method seems simple and obvious because it really is. Teachers have been doing this for ages. • The method is not unique. There are many aspects of this proposed method that are similar to existing methods. In spirit, we are all looking to address the same problem. • The difference is the perspective. If we think about the problem from the perspective of a client (e.g., Joe at DHS), what they look for in evaluation results currently are not the same as what we as researchers give them.
  • 14. Future Work • Integrate the proposed learning-based method to: – Grounded Evaluation – Long term effects (MILC)
  • 15. Thank you! rchang@uncc.edu http://www.viscenter.uncc.edu/~rchang
  • 16. The Classroom Analogy • Say you’re a math teacher in middle school, and you’re trying to decide which text book to use, the blue one or the red one. You can: – Ask your friends which book is better • Analogous to an “expert-based evaluation”. Problem is that the sample size is typically small, and the results difficult to replicate. – Ask your students which book they like • Analogous to subjective preferences. The issue here is that the students can prefer the blue text book because its blue. – Test which text book is more effective by giving the students tests.