SlideShare a Scribd company logo
micro tasks
s w i p e t o s t a r t
our goal: a standardized method for
benchmarking control room interfaces
• Tablet tool for data collection
• Can be used stand-alone or
• Linked to simulator
• Standardized data collection procedure
• Standardized method for question generation
• Standardized set of questions
• Database
micro tasks are for
 Evaluation
 HRA
 Training
Efficiently and objectively benchmark innovative
displays against conventional interfaces
IFE process overview
display on tablet
Example
IFE will conduct a micro task evaluation of this set-up
in December 2015 at a U.S. training simulator
Compare
IFE design concept for overview displays on tablet,
developed for the 2015 U.S. simulator study (screen 1 of 3)
example
innovative vs. conventional
mass balance
conventional innovative
Do the innovative displays lead to faster,
more reliable identifications and decisions?
mass balance
conventional innovative
Is performance (time and reliability) with innovative
displays at least as good as with conventional?
How to test performance benefits
of new interface solutions?
Scenario-based methods.
Observational or self-report.
Qualitative insights.
Relatively few data points.
Decontextualised.
Performance based.
Quantitative data.
Large amounts of data.
flavors of t&e methods
micro tasks are…
 Large number of questions
Related to systems, components, procedures, etc.
 Varying levels of difficulty
Including higher-level decision making
 Different display conditions
E.g. innovative vs. conventional displays
 We measure response time / accuracy
 Compare data between conditions
If needed, review eye tracking recordings
to understand anomalies
detection/decision tasks under time pressure
 Micro task tablet app linked to simulator
Tablet app receives signals from simulator
Tablet app can send signals / commands to simulator
 Dynamic scenarios
To cover monitoring / vigilance tasks
 We can now record operator actions
Can record operating of components, e.g. “start RCPs”
 New system for aggregating data
 New system for managing task lists
Makes it easier to set up and manage a study
new in 2015
example
Micro Task Interface Evaluation Method
4 operators
working in the
simulator
individually
(no communication)
We control
which displays
are available
(eg innovative, conventional)
Instructions:
”Please answer the
questions correctly,
but also as quickly
as possible.
It is very important
that you work as
fast as you can.”
Run 1
s w i p e t o s t a r t
How many condensate pumps
are running on turbine 31?
1 2 3
s w i p e t o
c o n t i n u e
“How many condensate pumps
are running on turbine 31?”
Average identification time
Conventional 8 sec
Innovative 6 sec
Is the subcooling margin sufficient?
Yes
s w i p e t o
c o n t i n u e
No
Which steam generators
are faulty?
SG-1
s w i p e t o
c o n t i n u e
SG-2 SG-3
Should safety injection
be stopped?
Yes
s w i p e t o
c o n t i n u e
No
What is the narrow-range level
in steam generator 1?
s w i p e t o
c o n t i n u e
%
1 2 3 4 5 6 7 8 9 0
37
up to 200 questions per hour
with 4 operators in the simulator working individually,
that means we can run up to 800 questions per hour
Tablet can trigger events in the simulator (e.g. start a tube leak)
and receive signal from the simulator (e.g. RCP-2 was started)
video
https://vimeo.com/131387407
results
data source
where does this data come from?
6840 data points
3420 response time measures
3420 accuracy measures
20 operators
5 hours simulator time
in total for the whole data collection
innovative displays are superior
HSI mean performance times
Current effect: F(2, 8,003 )=21,889, p=,00057
Vertical bars denote 0,95 confidence intervals
LSD (Innovative) OWD (Conventional) LSD and OWD
15
16
17
18
19
20
21
22
23
24
25
26
Performancetime(sec)
Operators were faster with innovative
displays than with conventional displays
innovative
conventional
HSI mean performance times
Current effect: F(2, 8,003 )=21,889, p=,00057
Vertical bars denote 0,95 confidence intervals
21
22
23
24
25
26
etime(sec)
Highly statistically
significant difference
Slightly more accurate with innovative
displays than conventional displays
Percentage of correct answers
innovative conventional innovative
Performance variability
conclusion
 As reliable as conventional displays
 Faster to read
innovative displays in the 2014 study are
summary
 Generate quantitative results about
performance benefits (time and reliability)
of digital/innovative CR interfaces
 Highly efficient, objective method
 Compare to existing reference data
 Qualitative insights via eye tracking
 Data directly supports HRA
 Highly customisable and precise
Generate exactly the data needed for the HRA
big picture
micro task tablet app = a mobile companion for
Human Factors / Human Reliability specialists
Micro task
tablet app
Standardized
question sets
Database
training
hra interface
design
gamification
evaluation of training
validation
benchmarking
within / across organisations
what next?
Benchmarking studies 2015 / 2016
benchmarking studies
(1) 2015 study at U.S. training simulator
analog interfaces vs IFE displays
(2) 2016 study of a partially digital control room
analog interfaces vs. 1990s-vintage overview
display vs IFE display
(3) 2016 study of a fully digital control room
Measure operator performance and reliability in a
fully digital control room and benchmark against
data from study 1 and 2
 Micro task database
 Micro tasks for training
Includes gamification approach
 Micro tasks for actions outside the
control room (field operator, mechanical, etc)
roadmap
 Micro tasks for team decisions
 Add secondary task capability
 Add workload measurement capability
roadmap
michael.hildebrandt@ife.no

More Related Content

PPTX
Poka Yoke with Matt Hansen at StatStuff
PDF
"Lean startups : what happens when you put hardware into the mix?" - Gal Shau...
PPT
Michael Roar Borlund & Christian Carlsen - Real Exploratory Testing, Now With...
PPTX
Process Capability: Step 4 (Normal Distributions)
PPTX
Promise 2011: Panel - "Practical Software Project Improvements using Actionab...
PPTX
Process Capability: Step 6 (Binomial)
PPTX
Calculating a Sample Size
PPTX
Introduction to Lean and Six Sigma by Matt Hansen at StatStuff (S02)
Poka Yoke with Matt Hansen at StatStuff
"Lean startups : what happens when you put hardware into the mix?" - Gal Shau...
Michael Roar Borlund & Christian Carlsen - Real Exploratory Testing, Now With...
Process Capability: Step 4 (Normal Distributions)
Promise 2011: Panel - "Practical Software Project Improvements using Actionab...
Process Capability: Step 6 (Binomial)
Calculating a Sample Size
Introduction to Lean and Six Sigma by Matt Hansen at StatStuff (S02)

What's hot (20)

PPTX
MSA – Planning & Conducting the MSA
PDF
How to get the most from your clinical outcome assessment (COA) measure - Tes...
PPTX
7 Deadly Wastes with Matt Hansen at StatStuff
PPTX
Work in Process with Matt Hansen at StatStuff
PPTX
Environmental Testing
PDF
Graham Freeburn - Make Your Testing Smarter - Know Your Context!
PPTX
Value Added with Matt Hansen at StatStuff
PPTX
Introduction to Lean with Matt Hansen at StatStuff
PPTX
Domains and data analytics
PPT
Michael Bolton - Heuristics: Solving Problems Rapidly
PPT
4 Of The 7 Problem Solving Tools
PPTX
Transfer Function with Matt Hansen at StatStuff
PDF
Experiment idea poster-p2
PDF
Software testing
PPTX
Change Acceleration Process (CAP) Model with Matt Hansen at StatStuff
PPTX
Defining the VOC and Defects
PPTX
Compiling Analysis Results
PPT
Shrini Kulkarni - Software Metrics - So Simple, Yet So Dangerous
PPTX
ALT Approaches for Reliability
PPTX
S.M.A.R.T & F.O.C.U.S Testing - Increasing the value provided by your testing...
MSA – Planning & Conducting the MSA
How to get the most from your clinical outcome assessment (COA) measure - Tes...
7 Deadly Wastes with Matt Hansen at StatStuff
Work in Process with Matt Hansen at StatStuff
Environmental Testing
Graham Freeburn - Make Your Testing Smarter - Know Your Context!
Value Added with Matt Hansen at StatStuff
Introduction to Lean with Matt Hansen at StatStuff
Domains and data analytics
Michael Bolton - Heuristics: Solving Problems Rapidly
4 Of The 7 Problem Solving Tools
Transfer Function with Matt Hansen at StatStuff
Experiment idea poster-p2
Software testing
Change Acceleration Process (CAP) Model with Matt Hansen at StatStuff
Defining the VOC and Defects
Compiling Analysis Results
Shrini Kulkarni - Software Metrics - So Simple, Yet So Dangerous
ALT Approaches for Reliability
S.M.A.R.T & F.O.C.U.S Testing - Increasing the value provided by your testing...
Ad

Viewers also liked (18)

PDF
Справка для "Бегунка"
PDF
Wild Wager
PDF
Global Fujitec Profile
PDF
MCDFeaturePagePdf451
PPTX
Russian vehicle park and fuel conumption
PDF
What We’ve Learned
PPTX
Bilgi Güvenliği
PDF
Risk SA Profile
PDF
2. Liberando el potencial humano economico a traves de micronutrientes
PDF
GOALQPCWhitepaperChangeManagement
PPTX
Tütün ve Zararları
PPTX
Rumah sehat
PPTX
Akses air ber
PDF
my work new
PPTX
Theories on the origin of the filipinos
PPTX
pmCareers Presentation
PPTX
Top 8 health education specialist resume samples
Справка для "Бегунка"
Wild Wager
Global Fujitec Profile
MCDFeaturePagePdf451
Russian vehicle park and fuel conumption
What We’ve Learned
Bilgi Güvenliği
Risk SA Profile
2. Liberando el potencial humano economico a traves de micronutrientes
GOALQPCWhitepaperChangeManagement
Tütün ve Zararları
Rumah sehat
Akses air ber
my work new
Theories on the origin of the filipinos
pmCareers Presentation
Top 8 health education specialist resume samples
Ad

Similar to Micro Task Interface Evaluation Method (20)

PPTX
Tablet tools and micro tasks
PDF
FlorenceAI: Reinventing Data Science at Humana
DOC
For an infinite queuing situation
PDF
Experiences with indicators
DOC
An assumption of learning curve theory is which of the following
PPTX
Training - What is Performance ?
PDF
big-book-of-data-science-2ndedition.pdf
DOC
You have been called in as a consultant to set up a kanban control system
PDF
How to Improve Quality and Efficiency Using Test Data Analytics
DOC
An advantage of a make to-stock process is which of the following
DOC
Which of the following is an input to the master production schedule (mps)
PPT
information retrival evaluation.ppt
DOC
Which of the following is a total measure of productivity
PPT
Workshop for newcomers
DOC
From an operational perspective, yield management is most effective under whi...
DOC
A company has actual unit demand for three consecutive years
DOC
A p chart to monitor process quality
PPT
Testwarez 2009 Use Proper Tool
PDF
OPS 571 HELP Expect Success /ops571help.com
PDF
Why Distributed Tracing is Essential for Performance and Reliability
Tablet tools and micro tasks
FlorenceAI: Reinventing Data Science at Humana
For an infinite queuing situation
Experiences with indicators
An assumption of learning curve theory is which of the following
Training - What is Performance ?
big-book-of-data-science-2ndedition.pdf
You have been called in as a consultant to set up a kanban control system
How to Improve Quality and Efficiency Using Test Data Analytics
An advantage of a make to-stock process is which of the following
Which of the following is an input to the master production schedule (mps)
information retrival evaluation.ppt
Which of the following is a total measure of productivity
Workshop for newcomers
From an operational perspective, yield management is most effective under whi...
A company has actual unit demand for three consecutive years
A p chart to monitor process quality
Testwarez 2009 Use Proper Tool
OPS 571 HELP Expect Success /ops571help.com
Why Distributed Tracing is Essential for Performance and Reliability

Recently uploaded (20)

PPT
Teaching material agriculture food technology
PDF
Encapsulation theory and applications.pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Assigned Numbers - 2025 - Bluetooth® Document
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PDF
A comparative analysis of optical character recognition models for extracting...
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Spectral efficient network and resource selection model in 5G networks
PPTX
Spectroscopy.pptx food analysis technology
PPTX
Big Data Technologies - Introduction.pptx
PDF
cuic standard and advanced reporting.pdf
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
Teaching material agriculture food technology
Encapsulation theory and applications.pdf
20250228 LYD VKU AI Blended-Learning.pptx
Assigned Numbers - 2025 - Bluetooth® Document
Reach Out and Touch Someone: Haptics and Empathic Computing
Agricultural_Statistics_at_a_Glance_2022_0.pdf
gpt5_lecture_notes_comprehensive_20250812015547.pdf
A comparative analysis of optical character recognition models for extracting...
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Chapter 3 Spatial Domain Image Processing.pdf
Per capita expenditure prediction using model stacking based on satellite ima...
Dropbox Q2 2025 Financial Results & Investor Presentation
Encapsulation_ Review paper, used for researhc scholars
Network Security Unit 5.pdf for BCA BBA.
Spectral efficient network and resource selection model in 5G networks
Spectroscopy.pptx food analysis technology
Big Data Technologies - Introduction.pptx
cuic standard and advanced reporting.pdf
“AI and Expert System Decision Support & Business Intelligence Systems”
Building Integrated photovoltaic BIPV_UPV.pdf

Micro Task Interface Evaluation Method

  • 1. micro tasks s w i p e t o s t a r t
  • 2. our goal: a standardized method for benchmarking control room interfaces • Tablet tool for data collection • Can be used stand-alone or • Linked to simulator • Standardized data collection procedure • Standardized method for question generation • Standardized set of questions • Database
  • 3. micro tasks are for  Evaluation  HRA  Training
  • 4. Efficiently and objectively benchmark innovative displays against conventional interfaces
  • 5. IFE process overview display on tablet Example IFE will conduct a micro task evaluation of this set-up in December 2015 at a U.S. training simulator
  • 7. IFE design concept for overview displays on tablet, developed for the 2015 U.S. simulator study (screen 1 of 3)
  • 9. mass balance conventional innovative Do the innovative displays lead to faster, more reliable identifications and decisions?
  • 10. mass balance conventional innovative Is performance (time and reliability) with innovative displays at least as good as with conventional?
  • 11. How to test performance benefits of new interface solutions?
  • 12. Scenario-based methods. Observational or self-report. Qualitative insights. Relatively few data points. Decontextualised. Performance based. Quantitative data. Large amounts of data. flavors of t&e methods
  • 14.  Large number of questions Related to systems, components, procedures, etc.  Varying levels of difficulty Including higher-level decision making  Different display conditions E.g. innovative vs. conventional displays  We measure response time / accuracy  Compare data between conditions If needed, review eye tracking recordings to understand anomalies detection/decision tasks under time pressure
  • 15.  Micro task tablet app linked to simulator Tablet app receives signals from simulator Tablet app can send signals / commands to simulator  Dynamic scenarios To cover monitoring / vigilance tasks  We can now record operator actions Can record operating of components, e.g. “start RCPs”  New system for aggregating data  New system for managing task lists Makes it easier to set up and manage a study new in 2015
  • 18. 4 operators working in the simulator individually (no communication)
  • 19. We control which displays are available (eg innovative, conventional)
  • 20. Instructions: ”Please answer the questions correctly, but also as quickly as possible. It is very important that you work as fast as you can.”
  • 21. Run 1 s w i p e t o s t a r t
  • 22. How many condensate pumps are running on turbine 31? 1 2 3 s w i p e t o c o n t i n u e
  • 23. “How many condensate pumps are running on turbine 31?” Average identification time Conventional 8 sec Innovative 6 sec
  • 24. Is the subcooling margin sufficient? Yes s w i p e t o c o n t i n u e No
  • 25. Which steam generators are faulty? SG-1 s w i p e t o c o n t i n u e SG-2 SG-3
  • 26. Should safety injection be stopped? Yes s w i p e t o c o n t i n u e No
  • 27. What is the narrow-range level in steam generator 1? s w i p e t o c o n t i n u e % 1 2 3 4 5 6 7 8 9 0 37
  • 28. up to 200 questions per hour with 4 operators in the simulator working individually, that means we can run up to 800 questions per hour
  • 29. Tablet can trigger events in the simulator (e.g. start a tube leak) and receive signal from the simulator (e.g. RCP-2 was started)
  • 32. data source where does this data come from?
  • 33. 6840 data points 3420 response time measures 3420 accuracy measures
  • 35. 5 hours simulator time in total for the whole data collection
  • 37. HSI mean performance times Current effect: F(2, 8,003 )=21,889, p=,00057 Vertical bars denote 0,95 confidence intervals LSD (Innovative) OWD (Conventional) LSD and OWD 15 16 17 18 19 20 21 22 23 24 25 26 Performancetime(sec) Operators were faster with innovative displays than with conventional displays innovative conventional HSI mean performance times Current effect: F(2, 8,003 )=21,889, p=,00057 Vertical bars denote 0,95 confidence intervals 21 22 23 24 25 26 etime(sec) Highly statistically significant difference
  • 38. Slightly more accurate with innovative displays than conventional displays Percentage of correct answers innovative conventional innovative
  • 41.  As reliable as conventional displays  Faster to read innovative displays in the 2014 study are
  • 43.  Generate quantitative results about performance benefits (time and reliability) of digital/innovative CR interfaces  Highly efficient, objective method  Compare to existing reference data  Qualitative insights via eye tracking  Data directly supports HRA  Highly customisable and precise Generate exactly the data needed for the HRA
  • 44. big picture micro task tablet app = a mobile companion for Human Factors / Human Reliability specialists
  • 45. Micro task tablet app Standardized question sets Database training hra interface design gamification evaluation of training validation benchmarking within / across organisations
  • 48. benchmarking studies (1) 2015 study at U.S. training simulator analog interfaces vs IFE displays (2) 2016 study of a partially digital control room analog interfaces vs. 1990s-vintage overview display vs IFE display (3) 2016 study of a fully digital control room Measure operator performance and reliability in a fully digital control room and benchmark against data from study 1 and 2
  • 49.  Micro task database  Micro tasks for training Includes gamification approach  Micro tasks for actions outside the control room (field operator, mechanical, etc) roadmap
  • 50.  Micro tasks for team decisions  Add secondary task capability  Add workload measurement capability roadmap

Editor's Notes

  • #37: Obviously this is not a general statement, but specific to the findings in this study
  • #38: There were 3 conditions: - Conventional displays only. Only the 30” operators screens were turned on (data in the middle) Innovative symbols only. Only the large screen was turned on (data on the left) Both. All screens were turned on, operators could choose which screens they use (data on the right)