SlideShare a Scribd company logo
Data Science
in the Cloud
Stefan Krawczyk
@stefkrawczyk
linkedin.com/in/skrawczyk
November 2016
InfoQ.com: News & Community Site
• Over 1,000,000 software developers, architects and CTOs read the site world-
wide every month
• 250,000 senior developers subscribe to our weekly newsletter
• Published in 4 languages (English, Chinese, Japanese and Brazilian
Portuguese)
• Post content from our QCon conferences
• 2 dedicated podcast channels: The InfoQ Podcast, with a focus on
Architecture and The Engineering Culture Podcast, with a focus on building
• 96 deep dives on innovative topics packed as downloadable emags and
minibooks
• Over 40 new content items per week
Watch the video with slide
synchronization on InfoQ.com!
https://www.infoq.com/presentations/
stitchfix-cloud
Purpose of QCon
- to empower software development by facilitating the spread of
knowledge and innovation
Strategy
- practitioner-driven conference designed for YOU: influencers of
change and innovation in your teams
- speakers and topics driving the evolution and innovation
- connecting and catalyzing the influencers and innovators
Highlights
- attended by more than 12,000 delegates since 2007
- held in 9 cities worldwide
Presented at QCon San Francisco
www.qconsf.com
Who are Data Scientists?
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Means: skills vary wildly
But they’re in
demand and expensive
“The Sexiest Job
of the 21st Century”
- HBR
https://hbr.org/2012/10/data-scientist-the-sexiest-job-of-the-21st-century
How many
Data Scientists do you have?
At Stitch Fix we have ~80
~85% have not done formal CS
But what do they do?
What is Stitch Fix?
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Data Science in the Cloud @StitchFix
Two Data Scientist facts:
1. Has AWS console access*.
2. End to end,
they’re responsible.
How do we enable this without
?
Make doing the right
thing the easy thing.
Fellow Collaborators
Horizontal team focused on Data Scientist Enablement
1. Eng. Skills
2. Important
3. What they work on
Let’s Start
Will Only Cover
1. Source of truth: S3 & Hive Metastore
2. Docker Enabled DS @ Stitch Fix
3. Scaling DS doing ML in the Cloud
Source of truth:
S3 & Hive Metastore
Want Everyone to Have Same View
A
B
This is Usually Nothing to Worry About
● OS handles correct access
● DB has ACID properties
A
B
This is Usually Nothing to Worry About
● OS handles correct access
● DB has ACID properties
● But it’s easy to outgrow these
options with a big data/team.
A
B
● Amazon’s Simple Storage Service
● Infinite* storage
● Can write, read, delete, BUT NOT append.
● Looks like a file system*:
○ URIs: my.bucket/path/to/files/file.txt
● Scales well
S3
* For all intents and purposes
● Hadoop service, that stores:
○ Schema
○ Partition information, e.g. date
○ Data location for a partition
Hive Metastore
● Hadoop service, that stores:
○ Schema
○ Partition information, e.g. date
○ Data location for a partition
Hive Metastore:
Hive Metastore
Partition Location
20161001 s3://bucket/sold_items/20161001
...
20161031 s3://bucket/sold_items/20161031
sold_items
Hive Metastore
● Replacing data in a partition
But if we’re not careful
● Replacing data in a partition
But if we’re not careful
But if we’re not careful
B
A
But if we’re not careful
● S3 is eventually
consistent
● These bugs are hard
to track down
A
B
● Use Hive Metastore to control partition source of truth
● Principles:
○ Never delete
○ Always write to a new place each time a partition changes
● Stitch Fix solution:
○ Use an inner directory → called Batch ID
Hive Metastore to the Rescue
Batch ID Pattern
Batch ID Pattern
Date Location
20161001 s3://bucket/sold_items/
20161001/20161002002334/
... ...
20161031 s3://bucket/sold_items/
20161031/20161101002256/
sold_items
● Overwriting a partition is just a matter of updating the location
Batch ID Pattern
Date Location
20161001 s3://bucket/sold_items/
20161001/20161002002334/
... ...
20161031 s3://bucket/sold_items/
20161031/20161101002256/
s3://bucket/sold_items/
20161031/20161102234252
sold_items
● Overwriting a partition is just a matter of updating the location
● To the user this is a hidden inner directory
Batch ID Pattern
Date Location
20161001 s3://bucket/sold_items/
20161001/20161002002334/
... ...
20161031 s3://bucket/sold_items/
20161031/20161101002256/
s3://bucket/sold_items/
20161031/20161102234252
sold_items
Enforce via API
Enforce via API
Python:
store_dataframe(df, dest_db, dest_table, partitions=[‘2016’])
df = load_dataframe(src_db, src_table, partitions=[‘2016’])
R:
sf_writer(data = result,
namespace = dest_db,
resource = dest_table,
partitions = c(as.integer(opt$ETL_DATE)))
sf_reader(namespace = src_db,
resource = src_table,
partitions = c(as.integer(opt$ETL_DATE)))
API for Data Scientists
● Full partition history
○ Can rollback
■ Data Scientists are less afraid of mistakes
○ Can create audit trails more easily
■ What data changed and when
○ Can anchor downstream consumers to a particular batch ID
Batch ID Pattern Benefits
Docker Enabled
DS @ Stitch Fix
Workstation Env. Mgmt. Scalability
Low Low
Medium Medium
High High
Ad hoc Infra: In the Beginning...
Workstation Env. Mgmt. Scalability
Low Low
Medium Medium
High High
Ad hoc Infra: Evolution I
Workstation Env. Mgmt. Scalability
Low Low
Medium Medium
High High
Ad hoc Infra: Evolution II
Workstation Env. Mgmt. Scalability
Low Low
Medium Medium
Low High
Ad hoc Infra: Evolution III
● Control of environment
○ Data Scientists don’t need to worry about env.
● Isolation
○ can host many docker containers on a single machine.
● Better host management
○ allowing central control of machine types.
Why Does Docker Lower Overhead?
Flotilla UI
● Has:
○ Our internal API libraries
○ Jupyter Notebook:
■ Pyspark
■ IPython
○ Python libs:
■ scikit, numpy, scipy, pandas, etc.
○ RStudio
○ R libs:
■ Dplyr, magrittr, ggplot2, lme4, BOOT, etc.
● Mounts User NFS
● User has terminal access to file system via Jupyter for git, pip, etc.
Our Docker Image
Docker Deployment
Docker Deployment
Docker Deployment
● Docker tightly integrates with the Linux Kernel.
○ Hypothesis:
■ Anything that makes uninterruptable calls to the kernel can:
● Break the ECS agent because the container doesn’t respond.
● Break isolation between containers.
■ E.g. Mounting NFS
● Docker Hub:
○ Switched to artifactory
Our Docker Problems So Far
Scaling
DS doing ML
in the Cloud
1. Data Latency
2. To Batch
or Not To Batch
3. What’s in a Model?
Data Latency
How much time do you spend waiting for data?
*This could be a laptop, a shared system, a batch process, etc.
Use Compression
*This could be a laptop, a shared system, a batch process, etc.
Use Compression - The Components
[ 1.3234543 0.23443434 … ]
[ 1 0 0 1 0 0 … 0 1 0 0
0 1 0 1 ...
… 1 0 1 1 ]
[ 1 0 0 1 0 0 … 0 1 0 0 ]
[ 1.3234543 0.23443434 … ]
{ 100: 0.56, … ,110: 0.65,
… , … , 999: 0.43 }
Use Compression - Python Comparison
Pickle: 60MB
Zlib+Pickle: 129KB
JSON: 15MB
Zlib+JSON: 55KB
Pickle: 3.1KB
Zlib+Pickle: 921B
JSON: 2.8KB
Zlib+JSON: 681B
Pickle: 2.6MB
Zlib+Pickle: 600KB
JSON: 769KB
Zlib+JSON: 139KB
[ 1.3234543 0.23443434 … ]
[ 1 0 0 1 0 0 … 0 1 0 0
0 1 0 1 ...
… 1 0 1 1 ]
[ 1 0 0 1 0 0 … 0 1 0 0 ]
[ 1.3234543 0.23443434 … ]
{ 100: 0.56, … ,110: 0.65,
… , … , 999: 0.43 }
● Naïve scheme of JSON + Zlib works well:
Observations
import json
import zlib
...
# compress
compressed = zlib.compress(json.dumps(value))
# decompress
original = json.loads(zlib.decompress(compressed))
● Naïve scheme of JSON + Zlib works well:
● Double vs Float: do you really need to store that much precision?
Observations
import json
import zlib
...
# compress
compressed = zlib.compress(json.dumps(value))
# decompress
original = json.loads(zlib.decompress(compressed))
● Naïve scheme of JSON + Zlib works well:
● Double vs Float: do you really need to store that much precision?
● For more inspiration look to columnar DBs and how they compress columns
Observations
import json
import zlib
...
# compress
compressed = zlib.compress(json.dumps(value))
# decompress
original = json.loads(zlib.decompress(compressed))
To Batch or Not To Batch:
When is batch inefficient?
● Online:
○ Computation occurs synchronously when needed.
● Streamed:
○ Computation is triggered by an event(s).
Online & Streamed Computation
Online & Streamed Computation
Very likely
you start with
a batch system
Online & Streamed Computation
● Do you need to recompute:
○ features for all users?
○ predicted results for all users?
Very likely
you start with
a batch system
Online & Streamed Computation
● Do you need to recompute:
○ features for all users?
○ predicted results for all users?
● Are you heavily dependent on your
ETL running every night?
Very likely
you start with
a batch system
Online & Streamed Computation
● Do you need to recompute:
○ features for all users?
○ predicted results for all users?
● Are you heavily dependent on your
ETL running every night?
● Online vs Streamed depends on in
house factors:
○ Number of models
○ How often they change
○ Cadence of output required
○ In house eng. expertise
○ etc.
Very likely
you start with
a batch system
Online & Streamed Computation
● Do you need to recompute:
○ features for all users?
○ predicted results for all users?
● Are you heavily dependent on your
ETL running every night?
● Online vs Streamed depends on in
house factors:
○ Number of models
○ How often they change
○ Cadence of output required
○ In house eng. expertise
○ etc.
Very likely
you start with
a batch system
We use online
system for
recommendations
Streamed Example
Streamed Example
Streamed Example
Streamed Example
● Dedicated infrastructure → More room on batch infrastructure
○ Hopefully $$$ savings
○ Hopefully less stressed Data Scientists
Online/Streaming Thoughts
● Dedicated infrastructure → More room on batch infrastructure
○ Hopefully $$$ savings
○ Hopefully less stressed Data Scientists
● Requires better software engineering practices
○ Code portability/reuse
○ Designing APIs/Tools Data Scientists will use
Online/Streaming Thoughts
● Dedicated infrastructure → More room on batch infrastructure
○ Hopefully $$$ savings
○ Hopefully less stressed Data Scientists
● Requires better software engineering practices
○ Code portability/reuse
○ Designing APIs/Tools Data Scientists will use
● Prototyping on AWS Lambda & Kinesis was surprisingly quick
○ Need to compile C libs on an amazon linux instance
Online/Streaming Thoughts
What’s in a Model?
Scaling model knowledge
Ever:
● Had someone leave and then nobody understands how they trained their
models?
Ever:
● Had someone leave and then nobody understands how they trained their
models?
○ Or you didn’t remember yourself?
Ever:
● Had someone leave and then nobody understands how they trained their
models?
○ Or you didn’t remember yourself?
● Had performance dip in models and you have trouble figuring out why?
Ever:
● Had someone leave and then nobody understands how they trained their
models?
○ Or you didn’t remember yourself?
● Had performance dip in models and you have trouble figuring out why?
○ Or not known what’s changed between model deployments?
Ever:
● Had someone leave and then nobody understands how they trained their
models?
○ Or you didn’t remember yourself?
● Had performance dip in models and you have trouble figuring out why?
○ Or not known what’s changed between model deployments?
● Wanted to compare model performance over time?
Ever:
● Had someone leave and then nobody understands how they trained their
models?
○ Or you didn’t remember yourself?
● Had performance dip in models and you have trouble figuring out why?
○ Or not known what’s changed between model deployments?
● Wanted to compare model performance over time?
● Wanted to train a model in R/Python/Spark and then deploy it a webserver?
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
○ NO!
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
How do you deal with
organizational drift?
Produce Model Artifacts
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
How do you deal with
organizational drift?
Produce Model Artifacts
Makes it easy to keep an
archive and track
changes over time
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
How do you deal with
organizational drift?
Produce Model Artifacts
Helps a lot with model
debugging & diagnosis!
Makes it easy to keep an
archive and track
changes over time
● Isn’t that just saving the coefficients/model values?
○ NO!
● Why?
How do you deal with
organizational drift?
Produce Model Artifacts
Helps a lot with model
debugging & diagnosis!
Makes it easy to keep an
archive and track
changes over time Can more easily use in
downstream processes
● Analogous to software libraries
● Packaging:
○ Zip/Jar file
Produce Model Artifacts
But all the above
seems complex?
We’re building APIs.
Fin; Questions?
@stefkrawczyk
Watch the video with slide
synchronization on InfoQ.com!
https://www.infoq.com/presentations/
stitchfix-cloud

More Related Content

PDF
Data Day Texas 2017: Scaling Data Science at Stitch Fix
PDF
Enabling Data Scientists to easily create and own Kafka Consumers
PDF
Managing your Black Friday Logs - Antonio Bonuccelli - Codemotion Rome 2018
ODP
Web-scale data processing: practical approaches for low-latency and batch
PPT
Building your own NSQL store
ODP
Big data nyu
PDF
Druid meetup 2018-03-13
PDF
Data pipelines from zero to solid
Data Day Texas 2017: Scaling Data Science at Stitch Fix
Enabling Data Scientists to easily create and own Kafka Consumers
Managing your Black Friday Logs - Antonio Bonuccelli - Codemotion Rome 2018
Web-scale data processing: practical approaches for low-latency and batch
Building your own NSQL store
Big data nyu
Druid meetup 2018-03-13
Data pipelines from zero to solid

What's hot (20)

PDF
SASI, Cassandra on the full text search ride - DuyHai Doan - Codemotion Milan...
PPTX
Building data pipelines
PPTX
SPARQL and Linked Data Benchmarking
PDF
Testing data streaming applications
PDF
Cassandra advanced data modeling
PDF
IOT with PostgreSQL
 
PPTX
Intro to Python Data Analysis in Wakari
PDF
Apache Spark vs rest of the world – Problems and Solutions by Arkadiusz Jachn...
PDF
Getting started with pandas
PPTX
A Developer’s View into Spark's Memory Model with Wenchen Fan
PDF
Dirty data? Clean it up! - Datapalooza Denver 2016
PDF
Test strategies for data processing pipelines, v2.0
PDF
Adding Complex Data to Spark Stack-(Neeraja Rentachintala, MapR)
PDF
Flink Forward Europe 2019 - Berlin
PDF
A look inside pandas design and development
PDF
A quick review of Python and Graph Databases
PDF
Tuning and Debugging in Apache Spark
PDF
Hadoop Pig: MapReduce the easy way!
PDF
Apache Spark Structured Streaming for Machine Learning - StrataConf 2016
PDF
Strata NYC 2015: Sketching Big Data with Spark: randomized algorithms for lar...
SASI, Cassandra on the full text search ride - DuyHai Doan - Codemotion Milan...
Building data pipelines
SPARQL and Linked Data Benchmarking
Testing data streaming applications
Cassandra advanced data modeling
IOT with PostgreSQL
 
Intro to Python Data Analysis in Wakari
Apache Spark vs rest of the world – Problems and Solutions by Arkadiusz Jachn...
Getting started with pandas
A Developer’s View into Spark's Memory Model with Wenchen Fan
Dirty data? Clean it up! - Datapalooza Denver 2016
Test strategies for data processing pipelines, v2.0
Adding Complex Data to Spark Stack-(Neeraja Rentachintala, MapR)
Flink Forward Europe 2019 - Berlin
A look inside pandas design and development
A quick review of Python and Graph Databases
Tuning and Debugging in Apache Spark
Hadoop Pig: MapReduce the easy way!
Apache Spark Structured Streaming for Machine Learning - StrataConf 2016
Strata NYC 2015: Sketching Big Data with Spark: randomized algorithms for lar...
Ad

Similar to Data Science in the Cloud @StitchFix (20)

PDF
Elastic Data Analytics Platform @Datadog
PDF
How to Develop and Operate Cloud First Data Platforms
PDF
MongoDB World 2019: Packing Up Your Data and Moving to MongoDB Atlas
PDF
OSMC 2018 | Learnings, patterns and Uber’s metrics platform M3, open sourced ...
PDF
Data Day Seattle 2017: Scaling Data Science at Stitch Fix
PDF
The Netflix Way to deal with Big Data Problems
PDF
Thinking DevOps in the Era of the Cloud - Demi Ben-Ari
PDF
Big data @ Hootsuite analtyics
PDF
Triangle Devops Meetup 10/2015
PPTX
Snowflake Automated Deployments / CI/CD Pipelines
PDF
Monitoring Big Data Systems Done "The Simple Way" - Codemotion Berlin 2017
PDF
Virtual training intro to InfluxDB - June 2021
PDF
CDC patterns in Apache Kafka®
PDF
Demi Ben-Ari - Monitoring Big Data Systems Done "The Simple Way" - Codemotion...
PDF
Monitoring Big Data Systems Done "The Simple Way" - Codemotion Milan 2017 - D...
PDF
Monitoring Big Data Systems "Done the simple way" - Demi Ben-Ari - Codemotion...
PDF
Monitoring Big Data Systems Done "The Simple Way" - Demi Ben-Ari - Codemotion...
PDF
Understanding Hadoop
PDF
OSMC 2023 | What’s new with Grafana Labs’s Open Source Observability stack by...
PDF
Cómo hemos implementado semántica de "Exactly Once" en nuestra base de datos ...
Elastic Data Analytics Platform @Datadog
How to Develop and Operate Cloud First Data Platforms
MongoDB World 2019: Packing Up Your Data and Moving to MongoDB Atlas
OSMC 2018 | Learnings, patterns and Uber’s metrics platform M3, open sourced ...
Data Day Seattle 2017: Scaling Data Science at Stitch Fix
The Netflix Way to deal with Big Data Problems
Thinking DevOps in the Era of the Cloud - Demi Ben-Ari
Big data @ Hootsuite analtyics
Triangle Devops Meetup 10/2015
Snowflake Automated Deployments / CI/CD Pipelines
Monitoring Big Data Systems Done "The Simple Way" - Codemotion Berlin 2017
Virtual training intro to InfluxDB - June 2021
CDC patterns in Apache Kafka®
Demi Ben-Ari - Monitoring Big Data Systems Done "The Simple Way" - Codemotion...
Monitoring Big Data Systems Done "The Simple Way" - Codemotion Milan 2017 - D...
Monitoring Big Data Systems "Done the simple way" - Demi Ben-Ari - Codemotion...
Monitoring Big Data Systems Done "The Simple Way" - Demi Ben-Ari - Codemotion...
Understanding Hadoop
OSMC 2023 | What’s new with Grafana Labs’s Open Source Observability stack by...
Cómo hemos implementado semántica de "Exactly Once" en nuestra base de datos ...
Ad

More from C4Media (20)

PDF
Streaming a Million Likes/Second: Real-Time Interactions on Live Video
PDF
Next Generation Client APIs in Envoy Mobile
PDF
Software Teams and Teamwork Trends Report Q1 2020
PDF
Understand the Trade-offs Using Compilers for Java Applications
PDF
Kafka Needs No Keeper
PDF
High Performing Teams Act Like Owners
PDF
Does Java Need Inline Types? What Project Valhalla Can Bring to Java
PDF
Service Meshes- The Ultimate Guide
PDF
Shifting Left with Cloud Native CI/CD
PDF
CI/CD for Machine Learning
PDF
Fault Tolerance at Speed
PDF
Architectures That Scale Deep - Regaining Control in Deep Systems
PDF
ML in the Browser: Interactive Experiences with Tensorflow.js
PDF
Build Your Own WebAssembly Compiler
PDF
User & Device Identity for Microservices @ Netflix Scale
PDF
Scaling Patterns for Netflix's Edge
PDF
Make Your Electron App Feel at Home Everywhere
PDF
The Talk You've Been Await-ing For
PDF
Future of Data Engineering
PDF
Automated Testing for Terraform, Docker, Packer, Kubernetes, and More
Streaming a Million Likes/Second: Real-Time Interactions on Live Video
Next Generation Client APIs in Envoy Mobile
Software Teams and Teamwork Trends Report Q1 2020
Understand the Trade-offs Using Compilers for Java Applications
Kafka Needs No Keeper
High Performing Teams Act Like Owners
Does Java Need Inline Types? What Project Valhalla Can Bring to Java
Service Meshes- The Ultimate Guide
Shifting Left with Cloud Native CI/CD
CI/CD for Machine Learning
Fault Tolerance at Speed
Architectures That Scale Deep - Regaining Control in Deep Systems
ML in the Browser: Interactive Experiences with Tensorflow.js
Build Your Own WebAssembly Compiler
User & Device Identity for Microservices @ Netflix Scale
Scaling Patterns for Netflix's Edge
Make Your Electron App Feel at Home Everywhere
The Talk You've Been Await-ing For
Future of Data Engineering
Automated Testing for Terraform, Docker, Packer, Kubernetes, and More

Recently uploaded (20)

PDF
Assigned Numbers - 2025 - Bluetooth® Document
PPTX
A Presentation on Touch Screen Technology
PDF
Approach and Philosophy of On baking technology
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PDF
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
PDF
1 - Historical Antecedents, Social Consideration.pdf
PDF
Microsoft Solutions Partner Drive Digital Transformation with D365.pdf
PDF
Univ-Connecticut-ChatGPT-Presentaion.pdf
PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PDF
project resource management chapter-09.pdf
PPTX
Tartificialntelligence_presentation.pptx
PPTX
Chapter 5: Probability Theory and Statistics
PDF
A novel scalable deep ensemble learning framework for big data classification...
PDF
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
PPTX
Group 1 Presentation -Planning and Decision Making .pptx
PDF
Mushroom cultivation and it's methods.pdf
Assigned Numbers - 2025 - Bluetooth® Document
A Presentation on Touch Screen Technology
Approach and Philosophy of On baking technology
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Unlocking AI with Model Context Protocol (MCP)
Agricultural_Statistics_at_a_Glance_2022_0.pdf
MIND Revenue Release Quarter 2 2025 Press Release
gpt5_lecture_notes_comprehensive_20250812015547.pdf
ENT215_Completing-a-large-scale-migration-and-modernization-with-AWS.pdf
1 - Historical Antecedents, Social Consideration.pdf
Microsoft Solutions Partner Drive Digital Transformation with D365.pdf
Univ-Connecticut-ChatGPT-Presentaion.pdf
Accuracy of neural networks in brain wave diagnosis of schizophrenia
project resource management chapter-09.pdf
Tartificialntelligence_presentation.pptx
Chapter 5: Probability Theory and Statistics
A novel scalable deep ensemble learning framework for big data classification...
From MVP to Full-Scale Product A Startup’s Software Journey.pdf
Group 1 Presentation -Planning and Decision Making .pptx
Mushroom cultivation and it's methods.pdf

Data Science in the Cloud @StitchFix

  • 1. Data Science in the Cloud Stefan Krawczyk @stefkrawczyk linkedin.com/in/skrawczyk November 2016
  • 2. InfoQ.com: News & Community Site • Over 1,000,000 software developers, architects and CTOs read the site world- wide every month • 250,000 senior developers subscribe to our weekly newsletter • Published in 4 languages (English, Chinese, Japanese and Brazilian Portuguese) • Post content from our QCon conferences • 2 dedicated podcast channels: The InfoQ Podcast, with a focus on Architecture and The Engineering Culture Podcast, with a focus on building • 96 deep dives on innovative topics packed as downloadable emags and minibooks • Over 40 new content items per week Watch the video with slide synchronization on InfoQ.com! https://www.infoq.com/presentations/ stitchfix-cloud
  • 3. Purpose of QCon - to empower software development by facilitating the spread of knowledge and innovation Strategy - practitioner-driven conference designed for YOU: influencers of change and innovation in your teams - speakers and topics driving the evolution and innovation - connecting and catalyzing the influencers and innovators Highlights - attended by more than 12,000 delegates since 2007 - held in 9 cities worldwide Presented at QCon San Francisco www.qconsf.com
  • 4. Who are Data Scientists?
  • 9. But they’re in demand and expensive
  • 10. “The Sexiest Job of the 21st Century” - HBR https://hbr.org/2012/10/data-scientist-the-sexiest-job-of-the-21st-century
  • 11. How many Data Scientists do you have?
  • 12. At Stitch Fix we have ~80
  • 13. ~85% have not done formal CS
  • 14. But what do they do?
  • 23. Two Data Scientist facts: 1. Has AWS console access*. 2. End to end, they’re responsible.
  • 24. How do we enable this without ?
  • 25. Make doing the right thing the easy thing.
  • 26. Fellow Collaborators Horizontal team focused on Data Scientist Enablement
  • 27. 1. Eng. Skills 2. Important 3. What they work on
  • 29. Will Only Cover 1. Source of truth: S3 & Hive Metastore 2. Docker Enabled DS @ Stitch Fix 3. Scaling DS doing ML in the Cloud
  • 30. Source of truth: S3 & Hive Metastore
  • 31. Want Everyone to Have Same View A B
  • 32. This is Usually Nothing to Worry About ● OS handles correct access ● DB has ACID properties A B
  • 33. This is Usually Nothing to Worry About ● OS handles correct access ● DB has ACID properties ● But it’s easy to outgrow these options with a big data/team. A B
  • 34. ● Amazon’s Simple Storage Service ● Infinite* storage ● Can write, read, delete, BUT NOT append. ● Looks like a file system*: ○ URIs: my.bucket/path/to/files/file.txt ● Scales well S3 * For all intents and purposes
  • 35. ● Hadoop service, that stores: ○ Schema ○ Partition information, e.g. date ○ Data location for a partition Hive Metastore
  • 36. ● Hadoop service, that stores: ○ Schema ○ Partition information, e.g. date ○ Data location for a partition Hive Metastore: Hive Metastore Partition Location 20161001 s3://bucket/sold_items/20161001 ... 20161031 s3://bucket/sold_items/20161031 sold_items
  • 38. ● Replacing data in a partition But if we’re not careful
  • 39. ● Replacing data in a partition But if we’re not careful
  • 40. But if we’re not careful B A
  • 41. But if we’re not careful ● S3 is eventually consistent ● These bugs are hard to track down A B
  • 42. ● Use Hive Metastore to control partition source of truth ● Principles: ○ Never delete ○ Always write to a new place each time a partition changes ● Stitch Fix solution: ○ Use an inner directory → called Batch ID Hive Metastore to the Rescue
  • 44. Batch ID Pattern Date Location 20161001 s3://bucket/sold_items/ 20161001/20161002002334/ ... ... 20161031 s3://bucket/sold_items/ 20161031/20161101002256/ sold_items
  • 45. ● Overwriting a partition is just a matter of updating the location Batch ID Pattern Date Location 20161001 s3://bucket/sold_items/ 20161001/20161002002334/ ... ... 20161031 s3://bucket/sold_items/ 20161031/20161101002256/ s3://bucket/sold_items/ 20161031/20161102234252 sold_items
  • 46. ● Overwriting a partition is just a matter of updating the location ● To the user this is a hidden inner directory Batch ID Pattern Date Location 20161001 s3://bucket/sold_items/ 20161001/20161002002334/ ... ... 20161031 s3://bucket/sold_items/ 20161031/20161101002256/ s3://bucket/sold_items/ 20161031/20161102234252 sold_items
  • 49. Python: store_dataframe(df, dest_db, dest_table, partitions=[‘2016’]) df = load_dataframe(src_db, src_table, partitions=[‘2016’]) R: sf_writer(data = result, namespace = dest_db, resource = dest_table, partitions = c(as.integer(opt$ETL_DATE))) sf_reader(namespace = src_db, resource = src_table, partitions = c(as.integer(opt$ETL_DATE))) API for Data Scientists
  • 50. ● Full partition history ○ Can rollback ■ Data Scientists are less afraid of mistakes ○ Can create audit trails more easily ■ What data changed and when ○ Can anchor downstream consumers to a particular batch ID Batch ID Pattern Benefits
  • 51. Docker Enabled DS @ Stitch Fix
  • 52. Workstation Env. Mgmt. Scalability Low Low Medium Medium High High Ad hoc Infra: In the Beginning...
  • 53. Workstation Env. Mgmt. Scalability Low Low Medium Medium High High Ad hoc Infra: Evolution I
  • 54. Workstation Env. Mgmt. Scalability Low Low Medium Medium High High Ad hoc Infra: Evolution II
  • 55. Workstation Env. Mgmt. Scalability Low Low Medium Medium Low High Ad hoc Infra: Evolution III
  • 56. ● Control of environment ○ Data Scientists don’t need to worry about env. ● Isolation ○ can host many docker containers on a single machine. ● Better host management ○ allowing central control of machine types. Why Does Docker Lower Overhead?
  • 58. ● Has: ○ Our internal API libraries ○ Jupyter Notebook: ■ Pyspark ■ IPython ○ Python libs: ■ scikit, numpy, scipy, pandas, etc. ○ RStudio ○ R libs: ■ Dplyr, magrittr, ggplot2, lme4, BOOT, etc. ● Mounts User NFS ● User has terminal access to file system via Jupyter for git, pip, etc. Our Docker Image
  • 62. ● Docker tightly integrates with the Linux Kernel. ○ Hypothesis: ■ Anything that makes uninterruptable calls to the kernel can: ● Break the ECS agent because the container doesn’t respond. ● Break isolation between containers. ■ E.g. Mounting NFS ● Docker Hub: ○ Switched to artifactory Our Docker Problems So Far
  • 64. 1. Data Latency 2. To Batch or Not To Batch 3. What’s in a Model?
  • 65. Data Latency How much time do you spend waiting for data?
  • 66. *This could be a laptop, a shared system, a batch process, etc.
  • 67. Use Compression *This could be a laptop, a shared system, a batch process, etc.
  • 68. Use Compression - The Components [ 1.3234543 0.23443434 … ] [ 1 0 0 1 0 0 … 0 1 0 0 0 1 0 1 ... … 1 0 1 1 ] [ 1 0 0 1 0 0 … 0 1 0 0 ] [ 1.3234543 0.23443434 … ] { 100: 0.56, … ,110: 0.65, … , … , 999: 0.43 }
  • 69. Use Compression - Python Comparison Pickle: 60MB Zlib+Pickle: 129KB JSON: 15MB Zlib+JSON: 55KB Pickle: 3.1KB Zlib+Pickle: 921B JSON: 2.8KB Zlib+JSON: 681B Pickle: 2.6MB Zlib+Pickle: 600KB JSON: 769KB Zlib+JSON: 139KB [ 1.3234543 0.23443434 … ] [ 1 0 0 1 0 0 … 0 1 0 0 0 1 0 1 ... … 1 0 1 1 ] [ 1 0 0 1 0 0 … 0 1 0 0 ] [ 1.3234543 0.23443434 … ] { 100: 0.56, … ,110: 0.65, … , … , 999: 0.43 }
  • 70. ● Naïve scheme of JSON + Zlib works well: Observations import json import zlib ... # compress compressed = zlib.compress(json.dumps(value)) # decompress original = json.loads(zlib.decompress(compressed))
  • 71. ● Naïve scheme of JSON + Zlib works well: ● Double vs Float: do you really need to store that much precision? Observations import json import zlib ... # compress compressed = zlib.compress(json.dumps(value)) # decompress original = json.loads(zlib.decompress(compressed))
  • 72. ● Naïve scheme of JSON + Zlib works well: ● Double vs Float: do you really need to store that much precision? ● For more inspiration look to columnar DBs and how they compress columns Observations import json import zlib ... # compress compressed = zlib.compress(json.dumps(value)) # decompress original = json.loads(zlib.decompress(compressed))
  • 73. To Batch or Not To Batch: When is batch inefficient?
  • 74. ● Online: ○ Computation occurs synchronously when needed. ● Streamed: ○ Computation is triggered by an event(s). Online & Streamed Computation
  • 75. Online & Streamed Computation Very likely you start with a batch system
  • 76. Online & Streamed Computation ● Do you need to recompute: ○ features for all users? ○ predicted results for all users? Very likely you start with a batch system
  • 77. Online & Streamed Computation ● Do you need to recompute: ○ features for all users? ○ predicted results for all users? ● Are you heavily dependent on your ETL running every night? Very likely you start with a batch system
  • 78. Online & Streamed Computation ● Do you need to recompute: ○ features for all users? ○ predicted results for all users? ● Are you heavily dependent on your ETL running every night? ● Online vs Streamed depends on in house factors: ○ Number of models ○ How often they change ○ Cadence of output required ○ In house eng. expertise ○ etc. Very likely you start with a batch system
  • 79. Online & Streamed Computation ● Do you need to recompute: ○ features for all users? ○ predicted results for all users? ● Are you heavily dependent on your ETL running every night? ● Online vs Streamed depends on in house factors: ○ Number of models ○ How often they change ○ Cadence of output required ○ In house eng. expertise ○ etc. Very likely you start with a batch system We use online system for recommendations
  • 84. ● Dedicated infrastructure → More room on batch infrastructure ○ Hopefully $$$ savings ○ Hopefully less stressed Data Scientists Online/Streaming Thoughts
  • 85. ● Dedicated infrastructure → More room on batch infrastructure ○ Hopefully $$$ savings ○ Hopefully less stressed Data Scientists ● Requires better software engineering practices ○ Code portability/reuse ○ Designing APIs/Tools Data Scientists will use Online/Streaming Thoughts
  • 86. ● Dedicated infrastructure → More room on batch infrastructure ○ Hopefully $$$ savings ○ Hopefully less stressed Data Scientists ● Requires better software engineering practices ○ Code portability/reuse ○ Designing APIs/Tools Data Scientists will use ● Prototyping on AWS Lambda & Kinesis was surprisingly quick ○ Need to compile C libs on an amazon linux instance Online/Streaming Thoughts
  • 87. What’s in a Model? Scaling model knowledge
  • 88. Ever: ● Had someone leave and then nobody understands how they trained their models?
  • 89. Ever: ● Had someone leave and then nobody understands how they trained their models? ○ Or you didn’t remember yourself?
  • 90. Ever: ● Had someone leave and then nobody understands how they trained their models? ○ Or you didn’t remember yourself? ● Had performance dip in models and you have trouble figuring out why?
  • 91. Ever: ● Had someone leave and then nobody understands how they trained their models? ○ Or you didn’t remember yourself? ● Had performance dip in models and you have trouble figuring out why? ○ Or not known what’s changed between model deployments?
  • 92. Ever: ● Had someone leave and then nobody understands how they trained their models? ○ Or you didn’t remember yourself? ● Had performance dip in models and you have trouble figuring out why? ○ Or not known what’s changed between model deployments? ● Wanted to compare model performance over time?
  • 93. Ever: ● Had someone leave and then nobody understands how they trained their models? ○ Or you didn’t remember yourself? ● Had performance dip in models and you have trouble figuring out why? ○ Or not known what’s changed between model deployments? ● Wanted to compare model performance over time? ● Wanted to train a model in R/Python/Spark and then deploy it a webserver?
  • 95. ● Isn’t that just saving the coefficients/model values? Produce Model Artifacts
  • 96. ● Isn’t that just saving the coefficients/model values? ○ NO! Produce Model Artifacts
  • 97. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? Produce Model Artifacts
  • 98. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? Produce Model Artifacts
  • 99. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? How do you deal with organizational drift? Produce Model Artifacts
  • 100. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? How do you deal with organizational drift? Produce Model Artifacts Makes it easy to keep an archive and track changes over time
  • 101. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? How do you deal with organizational drift? Produce Model Artifacts Helps a lot with model debugging & diagnosis! Makes it easy to keep an archive and track changes over time
  • 102. ● Isn’t that just saving the coefficients/model values? ○ NO! ● Why? How do you deal with organizational drift? Produce Model Artifacts Helps a lot with model debugging & diagnosis! Makes it easy to keep an archive and track changes over time Can more easily use in downstream processes
  • 103. ● Analogous to software libraries ● Packaging: ○ Zip/Jar file Produce Model Artifacts
  • 104. But all the above seems complex?
  • 107. Watch the video with slide synchronization on InfoQ.com! https://www.infoq.com/presentations/ stitchfix-cloud