SlideShare a Scribd company logo
Cisco and Big Data – Hadoop World 2011
Unified Fabric  optimized for Big Data infrastructures  with seamless integration with current data models Traditional Database RDBMS Storage SAN/NAS “ Big Data” Store And Analyze “ Big Data” Real-Time Capture, Read & Update  NoSQL Application Virtualized, Bare-Metal, Cloud Sensor Data Logs Social Media Click Streams Mobility Trends Event Data Cisco Unified Fabric
 
128 Nodes of UCS C200 M2, 1RU Rock-Mount Servers  4x2TB Drives, Dual Xeon 5670 @ 2.93GHz, 96GB RAM  The Setup Lab environment overview 16 Nodes of UCS C210 M2, 2RU Rack-Mount Servers 16xSFF Drives, Dual Xeon 5670 @2.93GHz, 96GB RAM
Unified Fabric L2 &/or L3 for SAN/NAS, RDBMS, UCS,  and Big Data L2/L3 Top of Rack Infrastructure … .. … .. Note: Two topologies were tested to examine the benefits of providing an integrated solution that can support multiple technologies, such as Traditional RDBMS, SAN/NAS, Virtualization, etc. N7K N7K N7K N3k N3k N3k N3k N3k N3k N7K N7K N7K N5k N5k N5k N5k … N2k … N2k UCS
Cluster Size Number of Data Nodes Data Model MapReduce functions Input Data Size Total starting dataset Characteristics of Data Node I/O, CPU, Memory, etc Data Locality in HDFS Ability to processes data where it already is located. Background Activity Number of Jobs running, type of jobs, Importing, exporting Networking Characteristics Availability, Buffering, 10GE vs 1GE, Oversubscription, Latency
A general characteristic of an optimally configured cluster is the ability to decrease job completion times by scaling out the nodes. Test results from ETL-like Workload (Yahoo Terasort) using 1TB data set.
The complexity of the functions used in Map and/or Reduce has a large impact on the type of job and network traffic. Note: Yahoo Terasort has a more balanced Map vs. Reduce functions and the same size of intermediate and final data (1TB Input, Shuffle and Output) Shakespeare WordCount has most of the processing in the Map Functions, smaller intermediate and even smaller final Data. (1TB Input, 10MB Shuffle, 1MB Output)
Network Graph of all Traffic received on an single node (80 Node run) Reducers Start Maps Finish Job Complete Maps Start These symbols represent a node sending traffic to HPC064 Note: Shortly after the Reducers start Map tasks are finishing and data is being shuffled to reducers As Maps completely finish the network is no loner used as Reducers have all the data they need to finish the job The red line is the total amount of traffic received by hpc064
Network Graph of all Traffic received on an single node (80 Node run) Output Data Replication Enabled Replication of 3 enabled (1 copy stored locally, 2 stored remotely) Each reduce output is replicated now, instead of just stored locally Note: If output replication is enabled, then the end of the terasort, must store additional copies. For a 1TB sort, 2TB will need to be replicated across the network.
Network Graph of all Traffic received on an single node (80 node run) Reducers Start Maps Finish Job Complete Maps Start The red line is the total amount of traffic received by hpc064 These symbols represent a node sending traffic to HPC064 Note: Due the the combination of the length of the Map phase and the reduced data set being shuffled, the network is being utilized throughout the job, but by a limited amount.
Given the same MapReduce Job, the larger the input dataset, the longer the job will take.  Note: It is important to note that as dataset sizes increase completion times may not scale linearly as many jobs can hit the ceiling of I/O and/or Compute power. Test results from ETL-like Workload (Yahoo Terasort) using varying data set sizes.
The I/O capacity, CPU and memory of the Data Node can have a direct impact on performance of a cluster. Note: A 2RU Server with 16 disks gives the node more storage, but trading off CPU per RU. On the other hand a 1RU server gives more CPU per rack.
Data Locality – The ability to process data where it is locally stored. Note: During the Map Phase, the JobTracker attempts to use data locality to schedule map tasks where the data is locally stored. This is not perfect and is dependent on a data nodes where the data is located. This is a consideration when choosing the replication factor. More replicas tend to create higher probability for data locality. Map Tasks: Initial spike for non-local data. Sometimes a task may be scheduled on a node that does not have the data available locally.
Hadoop clusters are generally multi-use. The effect of background use can effect any single jobs completion. Note: A given Cluster, is generally running may different types of Jobs, Importing into HDFS, Etc.  Example View of 24 Hour Cluster Use Large ETL Job Overlaps with medium and small ETL Jobs and many small BI Jobs (Blue lines are ETL Jobs and purple lines are BI Jobs) Importing Data into HDFS
The relative impact of various network characteristics on Hadoop clusters. Impact of Network Characteristics on Job Completion times
The failure of a networking device can affect multiple data nodes of a Hadoop cluster with a range of effects. Note: The tasks on affected nodes need to be rescheduled, schedule maintenance activities such as data rebalancing, increasing load on the cluster. Important to evaluate the overall availability of the system. Hadoop was designed with failure in mind, given any one node failure does not represent a huge issue, but network failures can span many nodes in the system causing rebalancing and decreased overall resources. Redundancy paths and load sharing schemes. General redundancy mechanisms can also increase bandwidth. Ease of management and consistent Operating System Main sources of outages can include human error. Ease of management and consistency are general best practices.
Several HDFS operations and phases of MapReduce jobs are very bursty in nature Note: The extent of bursts largely depend on the type of job (ETL vs. BI). Bursty phases can include replication of data (either importing into HDFS or output replication) and the output of the mappers during the shuffle phase. A network that cannot handle bursts effectively will drop packets, so optimal buffering is needed in network devices to absorb bursts. Optimal Buffering Given large enough incast, TCP will collapse at some point no matter how large the buffer Well studied by multiple universities Alternate solutions (Changing TCP behavior) proposed rather than Huge buffer switches ( http://simula. stanford .edu/sedcl/files/dc tcp -final.pdf )
Buffer being used during shuffle phase Buffer being used during output replication The buffer utilization is highest during the shuffle and output replication phases. Optimized buffer sizes are required to avoid packet loss leading to slower job completion times. Note: The Aggregation switch buffer remained flat as the bursts were absorbed at the Top of Rack layer
Buffer being used during shuffle phase Buffer being used during output replication Note: The Fabric Extender Buffer utilization was roughly equivalent to that of the N3k, but has 32MB of buffer vs. 9MB of N3k The buffer utilization is highest during the shuffle and output replication phases. Optimized buffer sizes are required to avoid packet loss leading to slower job completion times.
In a multi-use cluster described previously, multiple job types (ETL, BI, etc.) and Importing data into HDFS can be happening at the same time Note: Usage may vary depending on job scheduling options
In the largest workloads, multi-terabytes can be transmitted across the network Note: Data taken from multi-use workload (Multi-ETL + Multi-BI + HDFS Import).
Generally 1GE is being used largely due to the cost/performance trade-offs. Though 10GE can provide benefits depending on workload. Note: Multiple 1GE links can be bonded together to not only increase bandwidth, but increase bandwidth.
Moving from 1GE to 10GE actually lowers the buffer requirement on the switching layer.  Note: By moving to 10GE, the data node has a larger input to receive data lessening the need for buffers on the network as the total aggregate speed or amount of data does not increase substantially. This is due, in part, to limits of I/O and Compute capabilities
Generally network latency, while consistent latency being important, does not represent a significant factor for Hadoop Clusters. Note: There is a difference in network latency vs. application latency. Optimization in the application stack can decrease application latency that can potentially have a significant benefit.
For more information: www.cisco.com/go/bigdata
 

More Related Content

PDF
Hadoop World 2011: Next Generation Apache Hadoop MapReduce - Mohadev Konar, H...
PDF
MapReduce Improvements in MapR Hadoop
PDF
MapR M7: Providing an enterprise quality Apache HBase API
PPSX
MapReduce Scheduling Algorithms
PDF
Architectural Overview of MapR's Apache Hadoop Distribution
PPTX
Optimal Execution Of MapReduce Jobs In Cloud - Voices 2015
PPTX
10c introduction
PDF
Resource Aware Scheduling for Hadoop [Final Presentation]
Hadoop World 2011: Next Generation Apache Hadoop MapReduce - Mohadev Konar, H...
MapReduce Improvements in MapR Hadoop
MapR M7: Providing an enterprise quality Apache HBase API
MapReduce Scheduling Algorithms
Architectural Overview of MapR's Apache Hadoop Distribution
Optimal Execution Of MapReduce Jobs In Cloud - Voices 2015
10c introduction
Resource Aware Scheduling for Hadoop [Final Presentation]

What's hot (20)

PPTX
Introduction to Yarn
PDF
Hadoop scheduler
PPTX
Introduction to Yarn
PPTX
Inside MapR's M7
PDF
Design, Scale and Performance of MapR's Distribution for Hadoop
PPTX
Spark Overview and Performance Issues
PPTX
Distributed Processing Frameworks
PPTX
NYC Hadoop Meetup - MapR, Architecture, Philosophy and Applications
PDF
MapReduce and Hadoop
PDF
Hadoop data management
PPTX
KIISE:SIGDB Workshop presentation.
PPTX
Introduction to YARN and MapReduce 2
PPTX
The Future of Hadoop: MapR VP of Product Management, Tomer Shiran
PDF
Hadoop map reduce v2
PDF
The google MapReduce
PPT
Introduccion a Hadoop / Introduction to Hadoop
PPTX
Shift into High Gear: Dramatically Improve Hadoop & NoSQL Performance
PPTX
70a monitoring & troubleshooting
PDF
Hadoop Internals
PDF
An Introduction to MapReduce
Introduction to Yarn
Hadoop scheduler
Introduction to Yarn
Inside MapR's M7
Design, Scale and Performance of MapR's Distribution for Hadoop
Spark Overview and Performance Issues
Distributed Processing Frameworks
NYC Hadoop Meetup - MapR, Architecture, Philosophy and Applications
MapReduce and Hadoop
Hadoop data management
KIISE:SIGDB Workshop presentation.
Introduction to YARN and MapReduce 2
The Future of Hadoop: MapR VP of Product Management, Tomer Shiran
Hadoop map reduce v2
The google MapReduce
Introduccion a Hadoop / Introduction to Hadoop
Shift into High Gear: Dramatically Improve Hadoop & NoSQL Performance
70a monitoring & troubleshooting
Hadoop Internals
An Introduction to MapReduce
Ad

Viewers also liked (20)

PPTX
Apache HBase Performance Tuning
PPTX
Implementing the Hybrid Data Center
PDF
Wprowadzenie do technologii Big Data
PDF
Scalable midsize data center designs
PPT
Tweeting hadoop
PPT
Hadoop
PPTX
Monitoring and Analyzing Big Traffic Data of a Large-Scale Cellular Network w...
PDF
Hadoop Network Performance profile
PDF
Data Mining and Big Data Challenges and Research Opportunities
PPTX
Accelerating Apache Hadoop through High-Performance Networking and I/O Techno...
PDF
Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...
PPTX
Performing Network & Security Analytics with Hadoop
PPTX
ahepburn MDES PRES2 Production Tech Its only a Comic
PPTX
TRAFFIC DATA ANALYSIS USING HADOOP
ODP
Kkeithley ufonfs-gluster summit
PPTX
Network for the Large-scale Hadoop cluster at Yahoo! JAPAN
PPTX
Hands-On LinkedIn for Beginners
PPT
Chistesvarios8
PPTX
How to Add an App to Your LinkedIn Profile
PDF
ExcelCertificate18122014
Apache HBase Performance Tuning
Implementing the Hybrid Data Center
Wprowadzenie do technologii Big Data
Scalable midsize data center designs
Tweeting hadoop
Hadoop
Monitoring and Analyzing Big Traffic Data of a Large-Scale Cellular Network w...
Hadoop Network Performance profile
Data Mining and Big Data Challenges and Research Opportunities
Accelerating Apache Hadoop through High-Performance Networking and I/O Techno...
Deploying pNFS over Distributed File Storage w/ Jiffin Tony Thottan and Niels...
Performing Network & Security Analytics with Hadoop
ahepburn MDES PRES2 Production Tech Its only a Comic
TRAFFIC DATA ANALYSIS USING HADOOP
Kkeithley ufonfs-gluster summit
Network for the Large-scale Hadoop cluster at Yahoo! JAPAN
Hands-On LinkedIn for Beginners
Chistesvarios8
How to Add an App to Your LinkedIn Profile
ExcelCertificate18122014
Ad

Similar to Hadoop World 2011: Hadoop Network and Compute Architecture Considerations - Jacob Rapp, Cisco (20)

PPT
Seminar Presentation Hadoop
PPTX
Hadoop training-in-hyderabad
PPTX
Apache Hadoop India Summit 2011 talk "Hadoop Map-Reduce Programming & Best Pr...
PPTX
Exascale Capabl
PPT
Lecture Slide - Introduction to Hadoop, HDFS, MapR.ppt
PDF
getFamiliarWithHadoop
PDF
Big Data Architecture and Deployment
PDF
Cisco connect toronto 2015 big data sean mc keown
DOCX
Hadoop Research
PDF
Conference Paper: Universal Node: Towards a high-performance NFV environment
PDF
H04502048051
PPT
Hadoop - Introduction to HDFS
PPT
Hadoop online-training
PPTX
Programmable Exascale Supercomputer
PDF
Seminar_Report_hadoop
PPTX
Mapreduce is for Hadoop Ecosystem in Data Science
PDF
In datacenter performance analysis of a tensor processing unit
PDF
User-space Network Processing
PDF
MapReduce: Distributed Computing for Machine Learning
Seminar Presentation Hadoop
Hadoop training-in-hyderabad
Apache Hadoop India Summit 2011 talk "Hadoop Map-Reduce Programming & Best Pr...
Exascale Capabl
Lecture Slide - Introduction to Hadoop, HDFS, MapR.ppt
getFamiliarWithHadoop
Big Data Architecture and Deployment
Cisco connect toronto 2015 big data sean mc keown
Hadoop Research
Conference Paper: Universal Node: Towards a high-performance NFV environment
H04502048051
Hadoop - Introduction to HDFS
Hadoop online-training
Programmable Exascale Supercomputer
Seminar_Report_hadoop
Mapreduce is for Hadoop Ecosystem in Data Science
In datacenter performance analysis of a tensor processing unit
User-space Network Processing
MapReduce: Distributed Computing for Machine Learning

More from Cloudera, Inc. (20)

PPTX
Partner Briefing_January 25 (FINAL).pptx
PPTX
Cloudera Data Impact Awards 2021 - Finalists
PPTX
2020 Cloudera Data Impact Awards Finalists
PPTX
Edc event vienna presentation 1 oct 2019
PPTX
Machine Learning with Limited Labeled Data 4/3/19
PPTX
Data Driven With the Cloudera Modern Data Warehouse 3.19.19
PPTX
Introducing Cloudera DataFlow (CDF) 2.13.19
PPTX
Introducing Cloudera Data Science Workbench for HDP 2.12.19
PPTX
Shortening the Sales Cycle with a Modern Data Warehouse 1.30.19
PPTX
Leveraging the cloud for analytics and machine learning 1.29.19
PPTX
Modernizing the Legacy Data Warehouse – What, Why, and How 1.23.19
PPTX
Leveraging the Cloud for Big Data Analytics 12.11.18
PPTX
Modern Data Warehouse Fundamentals Part 3
PPTX
Modern Data Warehouse Fundamentals Part 2
PPTX
Modern Data Warehouse Fundamentals Part 1
PPTX
Extending Cloudera SDX beyond the Platform
PPTX
Federated Learning: ML with Privacy on the Edge 11.15.18
PPTX
Analyst Webinar: Doing a 180 on Customer 360
PPTX
Build a modern platform for anti-money laundering 9.19.18
PPTX
Introducing the data science sandbox as a service 8.30.18
Partner Briefing_January 25 (FINAL).pptx
Cloudera Data Impact Awards 2021 - Finalists
2020 Cloudera Data Impact Awards Finalists
Edc event vienna presentation 1 oct 2019
Machine Learning with Limited Labeled Data 4/3/19
Data Driven With the Cloudera Modern Data Warehouse 3.19.19
Introducing Cloudera DataFlow (CDF) 2.13.19
Introducing Cloudera Data Science Workbench for HDP 2.12.19
Shortening the Sales Cycle with a Modern Data Warehouse 1.30.19
Leveraging the cloud for analytics and machine learning 1.29.19
Modernizing the Legacy Data Warehouse – What, Why, and How 1.23.19
Leveraging the Cloud for Big Data Analytics 12.11.18
Modern Data Warehouse Fundamentals Part 3
Modern Data Warehouse Fundamentals Part 2
Modern Data Warehouse Fundamentals Part 1
Extending Cloudera SDX beyond the Platform
Federated Learning: ML with Privacy on the Edge 11.15.18
Analyst Webinar: Doing a 180 on Customer 360
Build a modern platform for anti-money laundering 9.19.18
Introducing the data science sandbox as a service 8.30.18

Recently uploaded (20)

PPT
Teaching material agriculture food technology
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
cuic standard and advanced reporting.pdf
PPTX
A Presentation on Artificial Intelligence
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Encapsulation theory and applications.pdf
PPTX
Cloud computing and distributed systems.
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Approach and Philosophy of On baking technology
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PDF
KodekX | Application Modernization Development
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
Teaching material agriculture food technology
“AI and Expert System Decision Support & Business Intelligence Systems”
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Network Security Unit 5.pdf for BCA BBA.
Reach Out and Touch Someone: Haptics and Empathic Computing
Unlocking AI with Model Context Protocol (MCP)
cuic standard and advanced reporting.pdf
A Presentation on Artificial Intelligence
Advanced methodologies resolving dimensionality complications for autism neur...
Building Integrated photovoltaic BIPV_UPV.pdf
Encapsulation theory and applications.pdf
Cloud computing and distributed systems.
Diabetes mellitus diagnosis method based random forest with bat algorithm
Approach and Philosophy of On baking technology
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
NewMind AI Weekly Chronicles - August'25 Week I
Chapter 3 Spatial Domain Image Processing.pdf
KodekX | Application Modernization Development
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
Dropbox Q2 2025 Financial Results & Investor Presentation

Hadoop World 2011: Hadoop Network and Compute Architecture Considerations - Jacob Rapp, Cisco

  • 1. Cisco and Big Data – Hadoop World 2011
  • 2. Unified Fabric optimized for Big Data infrastructures with seamless integration with current data models Traditional Database RDBMS Storage SAN/NAS “ Big Data” Store And Analyze “ Big Data” Real-Time Capture, Read & Update NoSQL Application Virtualized, Bare-Metal, Cloud Sensor Data Logs Social Media Click Streams Mobility Trends Event Data Cisco Unified Fabric
  • 3.  
  • 4. 128 Nodes of UCS C200 M2, 1RU Rock-Mount Servers 4x2TB Drives, Dual Xeon 5670 @ 2.93GHz, 96GB RAM The Setup Lab environment overview 16 Nodes of UCS C210 M2, 2RU Rack-Mount Servers 16xSFF Drives, Dual Xeon 5670 @2.93GHz, 96GB RAM
  • 5. Unified Fabric L2 &/or L3 for SAN/NAS, RDBMS, UCS, and Big Data L2/L3 Top of Rack Infrastructure … .. … .. Note: Two topologies were tested to examine the benefits of providing an integrated solution that can support multiple technologies, such as Traditional RDBMS, SAN/NAS, Virtualization, etc. N7K N7K N7K N3k N3k N3k N3k N3k N3k N7K N7K N7K N5k N5k N5k N5k … N2k … N2k UCS
  • 6. Cluster Size Number of Data Nodes Data Model MapReduce functions Input Data Size Total starting dataset Characteristics of Data Node I/O, CPU, Memory, etc Data Locality in HDFS Ability to processes data where it already is located. Background Activity Number of Jobs running, type of jobs, Importing, exporting Networking Characteristics Availability, Buffering, 10GE vs 1GE, Oversubscription, Latency
  • 7. A general characteristic of an optimally configured cluster is the ability to decrease job completion times by scaling out the nodes. Test results from ETL-like Workload (Yahoo Terasort) using 1TB data set.
  • 8. The complexity of the functions used in Map and/or Reduce has a large impact on the type of job and network traffic. Note: Yahoo Terasort has a more balanced Map vs. Reduce functions and the same size of intermediate and final data (1TB Input, Shuffle and Output) Shakespeare WordCount has most of the processing in the Map Functions, smaller intermediate and even smaller final Data. (1TB Input, 10MB Shuffle, 1MB Output)
  • 9. Network Graph of all Traffic received on an single node (80 Node run) Reducers Start Maps Finish Job Complete Maps Start These symbols represent a node sending traffic to HPC064 Note: Shortly after the Reducers start Map tasks are finishing and data is being shuffled to reducers As Maps completely finish the network is no loner used as Reducers have all the data they need to finish the job The red line is the total amount of traffic received by hpc064
  • 10. Network Graph of all Traffic received on an single node (80 Node run) Output Data Replication Enabled Replication of 3 enabled (1 copy stored locally, 2 stored remotely) Each reduce output is replicated now, instead of just stored locally Note: If output replication is enabled, then the end of the terasort, must store additional copies. For a 1TB sort, 2TB will need to be replicated across the network.
  • 11. Network Graph of all Traffic received on an single node (80 node run) Reducers Start Maps Finish Job Complete Maps Start The red line is the total amount of traffic received by hpc064 These symbols represent a node sending traffic to HPC064 Note: Due the the combination of the length of the Map phase and the reduced data set being shuffled, the network is being utilized throughout the job, but by a limited amount.
  • 12. Given the same MapReduce Job, the larger the input dataset, the longer the job will take. Note: It is important to note that as dataset sizes increase completion times may not scale linearly as many jobs can hit the ceiling of I/O and/or Compute power. Test results from ETL-like Workload (Yahoo Terasort) using varying data set sizes.
  • 13. The I/O capacity, CPU and memory of the Data Node can have a direct impact on performance of a cluster. Note: A 2RU Server with 16 disks gives the node more storage, but trading off CPU per RU. On the other hand a 1RU server gives more CPU per rack.
  • 14. Data Locality – The ability to process data where it is locally stored. Note: During the Map Phase, the JobTracker attempts to use data locality to schedule map tasks where the data is locally stored. This is not perfect and is dependent on a data nodes where the data is located. This is a consideration when choosing the replication factor. More replicas tend to create higher probability for data locality. Map Tasks: Initial spike for non-local data. Sometimes a task may be scheduled on a node that does not have the data available locally.
  • 15. Hadoop clusters are generally multi-use. The effect of background use can effect any single jobs completion. Note: A given Cluster, is generally running may different types of Jobs, Importing into HDFS, Etc. Example View of 24 Hour Cluster Use Large ETL Job Overlaps with medium and small ETL Jobs and many small BI Jobs (Blue lines are ETL Jobs and purple lines are BI Jobs) Importing Data into HDFS
  • 16. The relative impact of various network characteristics on Hadoop clusters. Impact of Network Characteristics on Job Completion times
  • 17. The failure of a networking device can affect multiple data nodes of a Hadoop cluster with a range of effects. Note: The tasks on affected nodes need to be rescheduled, schedule maintenance activities such as data rebalancing, increasing load on the cluster. Important to evaluate the overall availability of the system. Hadoop was designed with failure in mind, given any one node failure does not represent a huge issue, but network failures can span many nodes in the system causing rebalancing and decreased overall resources. Redundancy paths and load sharing schemes. General redundancy mechanisms can also increase bandwidth. Ease of management and consistent Operating System Main sources of outages can include human error. Ease of management and consistency are general best practices.
  • 18. Several HDFS operations and phases of MapReduce jobs are very bursty in nature Note: The extent of bursts largely depend on the type of job (ETL vs. BI). Bursty phases can include replication of data (either importing into HDFS or output replication) and the output of the mappers during the shuffle phase. A network that cannot handle bursts effectively will drop packets, so optimal buffering is needed in network devices to absorb bursts. Optimal Buffering Given large enough incast, TCP will collapse at some point no matter how large the buffer Well studied by multiple universities Alternate solutions (Changing TCP behavior) proposed rather than Huge buffer switches ( http://simula. stanford .edu/sedcl/files/dc tcp -final.pdf )
  • 19. Buffer being used during shuffle phase Buffer being used during output replication The buffer utilization is highest during the shuffle and output replication phases. Optimized buffer sizes are required to avoid packet loss leading to slower job completion times. Note: The Aggregation switch buffer remained flat as the bursts were absorbed at the Top of Rack layer
  • 20. Buffer being used during shuffle phase Buffer being used during output replication Note: The Fabric Extender Buffer utilization was roughly equivalent to that of the N3k, but has 32MB of buffer vs. 9MB of N3k The buffer utilization is highest during the shuffle and output replication phases. Optimized buffer sizes are required to avoid packet loss leading to slower job completion times.
  • 21. In a multi-use cluster described previously, multiple job types (ETL, BI, etc.) and Importing data into HDFS can be happening at the same time Note: Usage may vary depending on job scheduling options
  • 22. In the largest workloads, multi-terabytes can be transmitted across the network Note: Data taken from multi-use workload (Multi-ETL + Multi-BI + HDFS Import).
  • 23. Generally 1GE is being used largely due to the cost/performance trade-offs. Though 10GE can provide benefits depending on workload. Note: Multiple 1GE links can be bonded together to not only increase bandwidth, but increase bandwidth.
  • 24. Moving from 1GE to 10GE actually lowers the buffer requirement on the switching layer. Note: By moving to 10GE, the data node has a larger input to receive data lessening the need for buffers on the network as the total aggregate speed or amount of data does not increase substantially. This is due, in part, to limits of I/O and Compute capabilities
  • 25. Generally network latency, while consistent latency being important, does not represent a significant factor for Hadoop Clusters. Note: There is a difference in network latency vs. application latency. Optimization in the application stack can decrease application latency that can potentially have a significant benefit.
  • 26. For more information: www.cisco.com/go/bigdata
  • 27.