SlideShare a Scribd company logo
Improving experimentation
velocity via Multi-Armed Bandits
Dr Ilias Flaounas
Senior Data Scientist
Growth Hacking Meetup, Sydney, 20 June 2016
Improving experimentation velocity via Multi-Armed Bandits
http://www.nancydixonblog.com/2012/05/-why-knowledge-management-didnt-save-general-motors-addressing-complex-issues-
by-convening-conversat.html
Conversion rate
PDF•In a classic A/B we pick where to assign the next user randomly.
•In MAB we actively choose the cohort.
Pick black to exploit
Pick green (or red) to explore
Improving experimentation velocity via Multi-Armed Bandits
Improving experimentation velocity via Multi-Armed Bandits
Improving experimentation velocity via Multi-Armed Bandits
Improving experimentation velocity via Multi-Armed Bandits
Improving experimentation velocity via Multi-Armed Bandits
Win for variation “d”.
Win for variation “d” and estimation of p-values
Let’s run it for a bit longer… Again, win for variation “d”.
Classic A/B/C/D/E: ~2.5K samples
Multi-armed bandit: ~1K samples
60% Less samples
No winner after 1K iterations
Classic A/B/C: ~5K samples
Multi-armed bandit: ~1K samples
80% Less samples
No winner after 1K iterations
Classic A/B/C: ~2.8K samples
Multi-armed bandit: ~1K samples
64% Less samples
Win for variation “a”.
Classic A/B/C: ~1.8K samples
Multi-armed bandit: ~1K samples
45% Less samples
Disadvantages
• Reaching significance for
non-winning arms takes
longer
• Unclear stopping criteria -
App-specific heuristics
• Hard to order non-winning
arms and assess reliably
their impact
Advantages
• Reaching significance for
the winning arm is faster

• Best arm can change over
time
• There are no false
positives in the long term

Improving experimentation velocity via Multi-Armed Bandits
• How can we locate the
city of Bristol from
tweets?
• 10K candidate locations
organised in a 100x100
grid
• At every step we get
tweets from one
location and count the
number of mentions of
the word “Bristol”
• Challenge: find the target
in sub-linear time
complexity!
• Contextual bandits
can tackle this
problem
• We proposed the
KernelUCB, a non-
linear & contextual
flavour of MAB.
• The last few steps
of the algorithm
before it locates
Bristol.
Technical description: M. Valko, N. Korda, R. Munos, I. Flaounas, N. Cristianini, “Finite-
Time Analysis of Kernelised Contextual Bandits”, UAI, 2013.
Target is the red dot.
KernelUCB Matlab code: http://www.complacs.org/pmwiki.php/CompLACS/KernelUCB
KernelUCB
with RBF
kernel
converges
after ~300
iterations
(instead of
>>10K).
Improving experimentation velocity via Multi-Armed Bandits
Thank you!
Yes, we are hiring
Dr Ilias Flaounas
Senior Data Scientist

More Related Content

PDF
Multi-Armed Bandits:
 Intro, examples and tricks
PDF
Multi-Armed Bandit: an algorithmic perspective
PPTX
Website optimisation with Multi Armed Bandit algorithms
PPTX
Practical contextual bandits for business
PPTX
Byron Galbraith, Chief Data Scientist, Talla, at MLconf NYC 2017
PDF
Multi Armed Bandits
PDF
Claudia Vicol - Solving the multi armed bandit problem - Codemotion Amsterdam...
PDF
Bandit algorithms for website optimization - A summary
Multi-Armed Bandits:
 Intro, examples and tricks
Multi-Armed Bandit: an algorithmic perspective
Website optimisation with Multi Armed Bandit algorithms
Practical contextual bandits for business
Byron Galbraith, Chief Data Scientist, Talla, at MLconf NYC 2017
Multi Armed Bandits
Claudia Vicol - Solving the multi armed bandit problem - Codemotion Amsterdam...
Bandit algorithms for website optimization - A summary

Similar to Improving experimentation velocity via Multi-Armed Bandits (20)

PDF
Multi-Armed Bandit and Applications
PDF
Practical AI for Business: Bandit Algorithms
PDF
Contextual Bandit Survey
PDF
NON-STATIONARY BANDIT CHANGE DETECTION-BASED THOMPSON SAMPLING ALGORITHM
PDF
Multi-armed bandit by Joni Turunen
PPTX
Bandit algorithms
PDF
Bandit Algorithms
PDF
Artwork Personalization at Netflix
PDF
Minimize conversion loss in cro testing with multi armed bandits
 
PDF
Minimize conversion loss in cro testing with multi armed bandits (1)
 
PDF
Reinforcement Learning 2. Multi-armed Bandits
PDF
Multi-armed Bandits
PDF
John Maxwell, Data Scientist, Nordstrom at MLconf Seattle 2017
PDF
Scott Clark, Software Engineer, Yelp at MLconf SF
PPTX
multi-armed bandit
PDF
PDF
Optimal Learning for Fun and Profit with MOE
PPT
GAUSSIAN PRESENTATION (1).ppt
PPT
GAUSSIAN PRESENTATION.ppt
PDF
Personalized News Recommendation (Stream Data Based)
Multi-Armed Bandit and Applications
Practical AI for Business: Bandit Algorithms
Contextual Bandit Survey
NON-STATIONARY BANDIT CHANGE DETECTION-BASED THOMPSON SAMPLING ALGORITHM
Multi-armed bandit by Joni Turunen
Bandit algorithms
Bandit Algorithms
Artwork Personalization at Netflix
Minimize conversion loss in cro testing with multi armed bandits
 
Minimize conversion loss in cro testing with multi armed bandits (1)
 
Reinforcement Learning 2. Multi-armed Bandits
Multi-armed Bandits
John Maxwell, Data Scientist, Nordstrom at MLconf Seattle 2017
Scott Clark, Software Engineer, Yelp at MLconf SF
multi-armed bandit
Optimal Learning for Fun and Profit with MOE
GAUSSIAN PRESENTATION (1).ppt
GAUSSIAN PRESENTATION.ppt
Personalized News Recommendation (Stream Data Based)
Ad

More from Ilias Flaounas (9)

PDF
The story of the Product Growth team at Atlassian
PDF
On Storing Big Data
PDF
Readability and Linguistic Subjectivity of News
PDF
Detecting macro-patterns in the EU media sphere
PDF
Detecting Patterns in News Media Content
PDF
Celebrity Watch: Browsing News Content by Exploiting Social Intelligence
PDF
ECML/PKDD 2009: Found in translation
PDF
Data Science at Atlassian: 
The transition towards a data-driven organisation
PDF
Inference and validation of networks
The story of the Product Growth team at Atlassian
On Storing Big Data
Readability and Linguistic Subjectivity of News
Detecting macro-patterns in the EU media sphere
Detecting Patterns in News Media Content
Celebrity Watch: Browsing News Content by Exploiting Social Intelligence
ECML/PKDD 2009: Found in translation
Data Science at Atlassian: 
The transition towards a data-driven organisation
Inference and validation of networks
Ad

Recently uploaded (20)

PDF
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
PPTX
IB Computer Science - Internal Assessment.pptx
PPT
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
05. PRACTICAL GUIDE TO MICROSOFT EXCEL.pptx
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPTX
Moving the Public Sector (Government) to a Digital Adoption
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PDF
Lecture1 pattern recognition............
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PDF
Foundation of Data Science unit number two notes
PDF
Clinical guidelines as a resource for EBP(1).pdf
PDF
Mega Projects Data Mega Projects Data
PPTX
Global journeys: estimating international migration
PPTX
IBA_Chapter_11_Slides_Final_Accessible.pptx
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
IB Computer Science - Internal Assessment.pptx
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
Reliability_Chapter_ presentation 1221.5784
05. PRACTICAL GUIDE TO MICROSOFT EXCEL.pptx
Data_Analytics_and_PowerBI_Presentation.pptx
STUDY DESIGN details- Lt Col Maksud (21).pptx
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
Moving the Public Sector (Government) to a Digital Adoption
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
Lecture1 pattern recognition............
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
Acceptance and paychological effects of mandatory extra coach I classes.pptx
Foundation of Data Science unit number two notes
Clinical guidelines as a resource for EBP(1).pdf
Mega Projects Data Mega Projects Data
Global journeys: estimating international migration
IBA_Chapter_11_Slides_Final_Accessible.pptx

Improving experimentation velocity via Multi-Armed Bandits

  • 1. Improving experimentation velocity via Multi-Armed Bandits Dr Ilias Flaounas Senior Data Scientist Growth Hacking Meetup, Sydney, 20 June 2016
  • 4. Conversion rate PDF•In a classic A/B we pick where to assign the next user randomly. •In MAB we actively choose the cohort. Pick black to exploit Pick green (or red) to explore
  • 10. Win for variation “d”.
  • 11. Win for variation “d” and estimation of p-values
  • 12. Let’s run it for a bit longer… Again, win for variation “d”. Classic A/B/C/D/E: ~2.5K samples Multi-armed bandit: ~1K samples 60% Less samples
  • 13. No winner after 1K iterations Classic A/B/C: ~5K samples Multi-armed bandit: ~1K samples 80% Less samples
  • 14. No winner after 1K iterations Classic A/B/C: ~2.8K samples Multi-armed bandit: ~1K samples 64% Less samples
  • 15. Win for variation “a”. Classic A/B/C: ~1.8K samples Multi-armed bandit: ~1K samples 45% Less samples
  • 16. Disadvantages • Reaching significance for non-winning arms takes longer • Unclear stopping criteria - App-specific heuristics • Hard to order non-winning arms and assess reliably their impact Advantages • Reaching significance for the winning arm is faster
 • Best arm can change over time • There are no false positives in the long term

  • 18. • How can we locate the city of Bristol from tweets? • 10K candidate locations organised in a 100x100 grid • At every step we get tweets from one location and count the number of mentions of the word “Bristol” • Challenge: find the target in sub-linear time complexity!
  • 19. • Contextual bandits can tackle this problem • We proposed the KernelUCB, a non- linear & contextual flavour of MAB. • The last few steps of the algorithm before it locates Bristol. Technical description: M. Valko, N. Korda, R. Munos, I. Flaounas, N. Cristianini, “Finite- Time Analysis of Kernelised Contextual Bandits”, UAI, 2013.
  • 20. Target is the red dot. KernelUCB Matlab code: http://www.complacs.org/pmwiki.php/CompLACS/KernelUCB KernelUCB with RBF kernel converges after ~300 iterations (instead of >>10K).
  • 22. Thank you! Yes, we are hiring Dr Ilias Flaounas Senior Data Scientist