SlideShare a Scribd company logo
Hadoop in SIGMOD 20112011/5/20
PapersLCI: a social channel analysis platform for live customer intelligenceBistro data feed management systemApache hadoop goes realtime at FacebookNova: continuous Pig/Hadoop workflowsA Hadoop based distributed loading approach to parallel data warehousesA batch of PNUTS: experiences connecting cloud batch and serving systems
Papers (Continued)Turbocharging DBMS buffer pool using SSDsOnline reorganization in read optimized MMDBSAutomated partitioning design in parallel database systemsOracle database filesystemEmerging trends in the enterprise data analytics: connecting Hadoop and DB2 warehouseEfficient processing of data warehousing queries in a split execution environmentSQL server column store indexesAn analytic data engine for visualization in tableau
Apache Hadoop Goes Realtime at Facebook
Workload TypesFacebook MessagingHigh Write ThroughputLarge TablesData MigrationFacebook InsightsRealtime AnalyticsHigh Throughput IncrementsFacebook Metrics System (ODS)Automatic ShardingFast Reads of Recent Data and Table Scans
Why Hadoop & HBaseElasticityHigh write throughputEfficient and low-latency strong consistency semantics within a data centerEfficient random reads from diskHigh Availability and Disaster RecoveryFault IsolationAtomic read-modify-write primitivesRange ScansTolerance of network partitions within a single data centerZero Downtime in case of individual data center failureActive-active serving capability across different data centers
RealtimeHDFSHigh Availability - AvatarNodeHot Standby – AvatarNodeEnhancements to HDFS transaction loggingTransparent Failover: DAFS(client enhancement+ZooKeeper)HadoopRPC compatibilityBlock Availability: Placement Policya pluggable block placement policy
Realtime HDFS (Cont.)Performance Improvements for a Realtime WorkloadRPC TimeoutRecover File LeaseHDFS-appendrecoverLeaseReads from Local ReplicasNew FeaturesHDFS syncConcurrent Readers (last chunk of data)
Production HBaseACID Compliance (RWCC: Read Write Consistency Control)Atomicity (WALEdit)ConsistencyAvailability ImprovementsHBase Master RewriteRegion assignment in memory -> ZooKeeperOnline UpgradesDistributed Log SplittingPerformance ImprovementsCompactionRead Optimizations
Deployment and Operational ExperiencesTestingAuto Tesing ToolHBase VerifyMonitoring and ToolsHBCKMore metricsManual versus Automatic SplittingAdd new RegionServers, not region splittingDark Launch (灰度)Dashboards/ODS integrationBackups at the Application layerSchema ChangesImporting DataLzo & zipReducing Network IOMajor compaction
Nova: Continuous Pig/Hadoop Workflows
Nova OverviewScenariosIngesting and analyzing user behavior logs Building and updating a search index from a stream of crawled web pages Processing semi-structured data feedsTwo-layer programming model (Nova over Pig)Continuous processingIndependent schedulingCross-module optimizationManageability features
Abstract Workflow ModelWorkflowTwo kinds of vertices: tasks (processing steps) and channels (data containers)Edges connect tasks to channels and channels to tasksEdge annotations (all, new, B and Δ)Four common patterns of processingNon-incremental (template detection)Stateless incremental (shingling)Stateless incremental with lookup table (template tagging)Stateful incremental (de-duping)
Abstract Workflow Model (Cont.)Data and Update ModelBlocks: base blocks and delta blocksChannel functions: merge, chain and diffTask/Data InterfaceConsumption mode: all or newProduction mode: B or ΔWorkflow Programming and SchedulingData Compaction and Garbage Collection
Nova System Architecture
Efficient Processing of Data Warehousing Queries in a Split Execution Environment
IntroductionTwo approachesStarting with a parallel database system and adding some MapReduce featuresStarting with MapReduce and adding database system technologyHadoopDB follows the second of the approachesTwo heuristics for HadoopDB optimizationsDatabase systems can process data at a faster rate than Hadoop.Minimize the number of MapReduce jobs in SQL execution plan.
HadoopDBHadoopDB ArchitectureDatabase ConnectorData LoaderCatalogQuery InterfaceVectorWise/X100 Database (SIMD)  vs. PostgreSQLHadoopDB Query Executionselection, projection, and partial aggregation(Map and Combine)    database systemco-partitioned tablesMR for redistributing dataSideDB (a "database task done on the side").
Split Query ExecutionReferential PartitioningJoin in database engineLocal joinforeign-key  Referential PartitioningSplit MR/DB JoinsDirected join: one of the tables is already partitioned by the join key.Broadcast join: small table ought to be shipped to every node.Adding specialized joins to the MR framework  Map-side join.Tradeoffs: temporary table for join.Another type of join: MR redistributes data  Directed joinSplit MR/DB Semijoin like 'foreignKey IN (listOfValues)'Can be split into two MapReduce jobsSideDB to eliminate the first MapReduce job
Split Query Execution (Cont.)Post-join AggregationTwo MapReduce jobsHash-based partial aggregation   save significant I/OA similar technique is applied to TOP N selectionsPre-join AggregationFor MR based join.Group-by and join-key columns is smaller than the cardinality of the entire table.
A Query Plan in HadoopDB
PerformanceNo hash partition feature in Hive
Emerging Trends in the Enterprise Data Analytics: Connecting Hadoop and DB2 Warehouse
DB2 and Hadoop/Jaql Interactions
A HadoopBased Distributed Loading Approach to Parallel Data Warehouses
IntroductionWhy Hadoop for Teradata EDWMore disk space and space can be easily addedHDFS as a storageMapReduceDistributedHDFS blocks to Teradata EDW nodes assignment problemParameters: n blocks, k copies, m nodesGoal: to assign HDFS blocks to nodes evenly and minimize network traffic
Block Assignment Problem	HDFS file F on a cluster of P nodes (each node is uniquely identified with an integer i where 1 ≤ i ≤ P)    The problem is defined by:  assignment(X, Y, n,m, k, r) X is the set of n blocks (X = {1, . . . , n}) of FY is the set of m nodes running PDBMS (called PDBMS nodes) (Y⊆{1, . . . , P })k copies, m nodesr is the mapping recording the replicated block locations of each block.r(i) returns the set of nodes which has a copy of the block i.An assignment g from the blocks in X to the nodes in Y is denoted by a mapping from X = {1, . . . , n} to Y where g(i) = j (i ∈ X, j ∈ Y ) means that the block i is assigned to the node j.
Block Assignment Problem (Cont.)The problem is defined by:  assignment(X, Y, n,m, k, r)     An even assignment g is an assignment such that ∀ i ∈ Y ∀j ∈ Y| |{ x | ∀ 1 ≤ x ≤ n&&g(x) = i}| - |{y | ∀ 1 ≤ y ≤ n&&g(y) = j}| | ≤ 1. The cost of an assignment g is defined to be cost(g) = |{i | g(i) /∈r(i) ∀ 1 ≤ i ≤ n}|, which is the number of blocks assigned to remote nodes.We use |g| to denote the number of blocks assigned to local nodes by g. We have |g| = n - cost(g).The optimal assignment problem is to find an even assignment with the smallest cost.
OBA algorithm(X, Y, n,m, k, r)=({1, 2, 3}, {1, 2}, 3, 2, 1, {1 -> {1}, 2 -> {1}, 3 -> {2}})

More Related Content

PPTX
An introduction to Hadoop for large scale data analysis
PDF
Map Reduce data types and formats
PPTX
Hadoop MapReduce framework - Module 3
PPTX
Hadoop MapReduce joins
PPTX
Map reduce programming model to solve graph problems
PDF
PPTX
Sawmill - Integrating R and Large Data Clouds
PPTX
The strength of a spatial database
An introduction to Hadoop for large scale data analysis
Map Reduce data types and formats
Hadoop MapReduce framework - Module 3
Hadoop MapReduce joins
Map reduce programming model to solve graph problems
Sawmill - Integrating R and Large Data Clouds
The strength of a spatial database

What's hot (20)

PPTX
Hive and HiveQL - Module6
PPT
Taking Advantage of a Spatial Database with MapInfo Professional
PPTX
Database , 8 Query Optimization
PPTX
Spot db consistency checking and optimization in spatial database
PPTX
Using Spectrum on Demand from MapInfo Pro
PPTX
Qiu bosc2010
PDF
Introduction to the R Statistical Computing Environment
PDF
Computing Scientometrics in Large-Scale Academic Search Engines with MapReduce
PPTX
Hadoop Mapreduce joins
PPT
Comparing Distributed Indexing To Mapreduce or Not?
PDF
Hadoop Design Patterns
 
PPTX
CS267_Graph_Lab
PPT
Positional Data Organization and Compression in Web Inverted Indexes
PPT
MapReduce Over Lustre
PPTX
Incremental Export of Relational Database Contents into RDF Graphs
PDF
Aaa ped-6-Data manipulation: Data Files, and Data Cleaning & Preparation
PDF
Joey gonzalez, graph lab, m lconf 2013
PPTX
Geoprocessing
PPTX
Advance Hive, NoSQL Database (HBase) - Module 7
Hive and HiveQL - Module6
Taking Advantage of a Spatial Database with MapInfo Professional
Database , 8 Query Optimization
Spot db consistency checking and optimization in spatial database
Using Spectrum on Demand from MapInfo Pro
Qiu bosc2010
Introduction to the R Statistical Computing Environment
Computing Scientometrics in Large-Scale Academic Search Engines with MapReduce
Hadoop Mapreduce joins
Comparing Distributed Indexing To Mapreduce or Not?
Hadoop Design Patterns
 
CS267_Graph_Lab
Positional Data Organization and Compression in Web Inverted Indexes
MapReduce Over Lustre
Incremental Export of Relational Database Contents into RDF Graphs
Aaa ped-6-Data manipulation: Data Files, and Data Cleaning & Preparation
Joey gonzalez, graph lab, m lconf 2013
Geoprocessing
Advance Hive, NoSQL Database (HBase) - Module 7
Ad

Viewers also liked (7)

PPT
C++编程实践
PPTX
Qii12
PPT
La Comunicació Amb Els Nostres Fills I Filles
PPT
Usability
PPT
eLearning - mLearning
PPTX
Intro To Tendenci - JJ Lassberg
KEY
How to grow a profitable assocation
C++编程实践
Qii12
La Comunicació Amb Els Nostres Fills I Filles
Usability
eLearning - mLearning
Intro To Tendenci - JJ Lassberg
How to grow a profitable assocation
Ad

Similar to Hadoop in sigmod 2011 (20)

PPT
Hive @ Hadoop day seattle_2010
PDF
Quick dive into the big data pool without drowning - Demi Ben-Ari @ Panorays
PPT
Eric Baldeschwieler Keynote from Storage Developers Conference
PPTX
عصر کلان داده، چرا و چگونه؟
PPT
Big Data
PPTX
Hadoop introduction , Why and What is Hadoop ?
PPTX
Hadoop & distributed cloud computing
PDF
Big data processing systems research
PDF
Hadoop on Azure, Blue elephants
PPTX
BigData
PDF
Asd 2015
PPTX
Bigdata
PPT
Nextag talk
PDF
Hadoop programming
PDF
Introduction to Hadoop
PPTX
Introduction of Big data, NoSQL & Hadoop
PPTX
Bigdata and Hadoop
PPTX
Coding serbia
PPT
Finding the needles in the haystack. An Overview of Analyzing Big Data with H...
PPTX
Real time hadoop + mapreduce intro
Hive @ Hadoop day seattle_2010
Quick dive into the big data pool without drowning - Demi Ben-Ari @ Panorays
Eric Baldeschwieler Keynote from Storage Developers Conference
عصر کلان داده، چرا و چگونه؟
Big Data
Hadoop introduction , Why and What is Hadoop ?
Hadoop & distributed cloud computing
Big data processing systems research
Hadoop on Azure, Blue elephants
BigData
Asd 2015
Bigdata
Nextag talk
Hadoop programming
Introduction to Hadoop
Introduction of Big data, NoSQL & Hadoop
Bigdata and Hadoop
Coding serbia
Finding the needles in the haystack. An Overview of Analyzing Big Data with H...
Real time hadoop + mapreduce intro

Recently uploaded (20)

PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PPTX
MYSQL Presentation for SQL database connectivity
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
Electronic commerce courselecture one. Pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Encapsulation theory and applications.pdf
PDF
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
CIFDAQ's Market Insight: SEC Turns Pro Crypto
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PDF
Empathic Computing: Creating Shared Understanding
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PPT
Teaching material agriculture food technology
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
Digital-Transformation-Roadmap-for-Companies.pptx
MYSQL Presentation for SQL database connectivity
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Electronic commerce courselecture one. Pdf
20250228 LYD VKU AI Blended-Learning.pptx
Encapsulation theory and applications.pdf
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
Per capita expenditure prediction using model stacking based on satellite ima...
Network Security Unit 5.pdf for BCA BBA.
Agricultural_Statistics_at_a_Glance_2022_0.pdf
CIFDAQ's Market Insight: SEC Turns Pro Crypto
Dropbox Q2 2025 Financial Results & Investor Presentation
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Chapter 3 Spatial Domain Image Processing.pdf
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
Empathic Computing: Creating Shared Understanding
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Teaching material agriculture food technology

Hadoop in sigmod 2011

  • 1. Hadoop in SIGMOD 20112011/5/20
  • 2. PapersLCI: a social channel analysis platform for live customer intelligenceBistro data feed management systemApache hadoop goes realtime at FacebookNova: continuous Pig/Hadoop workflowsA Hadoop based distributed loading approach to parallel data warehousesA batch of PNUTS: experiences connecting cloud batch and serving systems
  • 3. Papers (Continued)Turbocharging DBMS buffer pool using SSDsOnline reorganization in read optimized MMDBSAutomated partitioning design in parallel database systemsOracle database filesystemEmerging trends in the enterprise data analytics: connecting Hadoop and DB2 warehouseEfficient processing of data warehousing queries in a split execution environmentSQL server column store indexesAn analytic data engine for visualization in tableau
  • 4. Apache Hadoop Goes Realtime at Facebook
  • 5. Workload TypesFacebook MessagingHigh Write ThroughputLarge TablesData MigrationFacebook InsightsRealtime AnalyticsHigh Throughput IncrementsFacebook Metrics System (ODS)Automatic ShardingFast Reads of Recent Data and Table Scans
  • 6. Why Hadoop & HBaseElasticityHigh write throughputEfficient and low-latency strong consistency semantics within a data centerEfficient random reads from diskHigh Availability and Disaster RecoveryFault IsolationAtomic read-modify-write primitivesRange ScansTolerance of network partitions within a single data centerZero Downtime in case of individual data center failureActive-active serving capability across different data centers
  • 7. RealtimeHDFSHigh Availability - AvatarNodeHot Standby – AvatarNodeEnhancements to HDFS transaction loggingTransparent Failover: DAFS(client enhancement+ZooKeeper)HadoopRPC compatibilityBlock Availability: Placement Policya pluggable block placement policy
  • 8. Realtime HDFS (Cont.)Performance Improvements for a Realtime WorkloadRPC TimeoutRecover File LeaseHDFS-appendrecoverLeaseReads from Local ReplicasNew FeaturesHDFS syncConcurrent Readers (last chunk of data)
  • 9. Production HBaseACID Compliance (RWCC: Read Write Consistency Control)Atomicity (WALEdit)ConsistencyAvailability ImprovementsHBase Master RewriteRegion assignment in memory -> ZooKeeperOnline UpgradesDistributed Log SplittingPerformance ImprovementsCompactionRead Optimizations
  • 10. Deployment and Operational ExperiencesTestingAuto Tesing ToolHBase VerifyMonitoring and ToolsHBCKMore metricsManual versus Automatic SplittingAdd new RegionServers, not region splittingDark Launch (灰度)Dashboards/ODS integrationBackups at the Application layerSchema ChangesImporting DataLzo & zipReducing Network IOMajor compaction
  • 12. Nova OverviewScenariosIngesting and analyzing user behavior logs Building and updating a search index from a stream of crawled web pages Processing semi-structured data feedsTwo-layer programming model (Nova over Pig)Continuous processingIndependent schedulingCross-module optimizationManageability features
  • 13. Abstract Workflow ModelWorkflowTwo kinds of vertices: tasks (processing steps) and channels (data containers)Edges connect tasks to channels and channels to tasksEdge annotations (all, new, B and Δ)Four common patterns of processingNon-incremental (template detection)Stateless incremental (shingling)Stateless incremental with lookup table (template tagging)Stateful incremental (de-duping)
  • 14. Abstract Workflow Model (Cont.)Data and Update ModelBlocks: base blocks and delta blocksChannel functions: merge, chain and diffTask/Data InterfaceConsumption mode: all or newProduction mode: B or ΔWorkflow Programming and SchedulingData Compaction and Garbage Collection
  • 16. Efficient Processing of Data Warehousing Queries in a Split Execution Environment
  • 17. IntroductionTwo approachesStarting with a parallel database system and adding some MapReduce featuresStarting with MapReduce and adding database system technologyHadoopDB follows the second of the approachesTwo heuristics for HadoopDB optimizationsDatabase systems can process data at a faster rate than Hadoop.Minimize the number of MapReduce jobs in SQL execution plan.
  • 18. HadoopDBHadoopDB ArchitectureDatabase ConnectorData LoaderCatalogQuery InterfaceVectorWise/X100 Database (SIMD) vs. PostgreSQLHadoopDB Query Executionselection, projection, and partial aggregation(Map and Combine)  database systemco-partitioned tablesMR for redistributing dataSideDB (a "database task done on the side").
  • 19. Split Query ExecutionReferential PartitioningJoin in database engineLocal joinforeign-key  Referential PartitioningSplit MR/DB JoinsDirected join: one of the tables is already partitioned by the join key.Broadcast join: small table ought to be shipped to every node.Adding specialized joins to the MR framework  Map-side join.Tradeoffs: temporary table for join.Another type of join: MR redistributes data  Directed joinSplit MR/DB Semijoin like 'foreignKey IN (listOfValues)'Can be split into two MapReduce jobsSideDB to eliminate the first MapReduce job
  • 20. Split Query Execution (Cont.)Post-join AggregationTwo MapReduce jobsHash-based partial aggregation  save significant I/OA similar technique is applied to TOP N selectionsPre-join AggregationFor MR based join.Group-by and join-key columns is smaller than the cardinality of the entire table.
  • 21. A Query Plan in HadoopDB
  • 23. Emerging Trends in the Enterprise Data Analytics: Connecting Hadoop and DB2 Warehouse
  • 24. DB2 and Hadoop/Jaql Interactions
  • 25. A HadoopBased Distributed Loading Approach to Parallel Data Warehouses
  • 26. IntroductionWhy Hadoop for Teradata EDWMore disk space and space can be easily addedHDFS as a storageMapReduceDistributedHDFS blocks to Teradata EDW nodes assignment problemParameters: n blocks, k copies, m nodesGoal: to assign HDFS blocks to nodes evenly and minimize network traffic
  • 27. Block Assignment Problem HDFS file F on a cluster of P nodes (each node is uniquely identified with an integer i where 1 ≤ i ≤ P) The problem is defined by: assignment(X, Y, n,m, k, r) X is the set of n blocks (X = {1, . . . , n}) of FY is the set of m nodes running PDBMS (called PDBMS nodes) (Y⊆{1, . . . , P })k copies, m nodesr is the mapping recording the replicated block locations of each block.r(i) returns the set of nodes which has a copy of the block i.An assignment g from the blocks in X to the nodes in Y is denoted by a mapping from X = {1, . . . , n} to Y where g(i) = j (i ∈ X, j ∈ Y ) means that the block i is assigned to the node j.
  • 28. Block Assignment Problem (Cont.)The problem is defined by: assignment(X, Y, n,m, k, r) An even assignment g is an assignment such that ∀ i ∈ Y ∀j ∈ Y| |{ x | ∀ 1 ≤ x ≤ n&&g(x) = i}| - |{y | ∀ 1 ≤ y ≤ n&&g(y) = j}| | ≤ 1. The cost of an assignment g is defined to be cost(g) = |{i | g(i) /∈r(i) ∀ 1 ≤ i ≤ n}|, which is the number of blocks assigned to remote nodes.We use |g| to denote the number of blocks assigned to local nodes by g. We have |g| = n - cost(g).The optimal assignment problem is to find an even assignment with the smallest cost.
  • 29. OBA algorithm(X, Y, n,m, k, r)=({1, 2, 3}, {1, 2}, 3, 2, 1, {1 -> {1}, 2 -> {1}, 3 -> {2}})