SlideShare a Scribd company logo
Rgk cluster computing project
FAULT TOLERANCE IN
CLUSTER COMPUTING
Guided By- Submitted By-
Mr. Ankush Agrawal Ravindra Pratap Singh
Mr. Praveen Rai Garima Kaushik
Kamini Saraswat
OUTLINE
 Introduction
Purpose
Requirements
Advantages Of Linux
 Objective
 Sub-Objective
 Research Gap
 Basic MPI Commands
Message Passing Interface
Working Strategy
Graphical Representation
INTRODUCTION
What Is Cluster …???
A cluster is a set of connected computers that work
together so that it can be viewed as a single system. It
works on master slave connection.
 What Is Cluster Computing…???
A cluster computing is also known as HPC as it is
used to solve the large problems in less time compared
with other techniques. HPC may include Parallel,
Cluster, Grid, Cloud and Green computing.
CONTINUE...
What Is Fault…???
A fault is any error or unwanted condition that
may arise in a system due to which our system will stop
its execution. It may be a natural or man-made types.
 What Is Fault Tolerance…???
A fault tolerance is an ability by which we will
tolerate some type of faults so that we will get the
correct final outcome. Eg. Faulty processor etc.
PUPOSE
The purpose of cluster technology is to eliminate single
points of failure. When availability of data is your
paramount consideration, clustering is ideal. Using a
cluster we can avoids all of these single points of
failure:
 Network card failure
 Processor failure
 Motherboard failure
REQUIREMENTS
Software Environment
Operating system- Ubuntu 10.0.4 LTS
MPI_Ch2 Package
Open_MPI
Libshem-dev
Libmpich2-dev
ADVANTAGES OF USING
LINUX
The following are some advantages of using Linux:
Linux is readily available on the Internet and can be
downloaded without cost.
It is easy to fix bugs and improve system performance.
Users can develop or fine-tune hardware drivers which
can easily be made available to other users.
The most important advantage of using linux is that it
creates a several copies of one processor which helps in
enhancing the performance of a system.
OBJECTIVE
We are working on linux operating system & on a
communication patterns of clusters using MPI.
Our aim is to find faults, and to recover those faults
which are causing unexpected behaviours (error , bugs
etc.).
MESSAGE PASSING
INTERFACE(MPI)
The generic form of message passing in the parallel
computing is the Message Passing Interface.
It is used as a medium of communication among the
nodes.
In message passing, data is moved from address space
of one to that of other by mean of cooperative
operation such as send/receive pair.
BASIC MPI
ROUTINS/COMMANDS
For comunication among different processes some routines
are used which are-
MPI_Send, to send a message to another process.
MPI_Recv, to receive a message from another process.
MPI_Gather, MPI_Gatherv, to gather data from
participating processes into a single structure.
 MPI Comm size() – Number of MPI processes.
 MPI Comm rank() – Internal process number.
 MPI Get processor name() – External processor name.
CONTINUE…
 MPI_Scatter, MPI_Scatter, to break a structure into
portions and distribute those portions to other processes.
 MPI_Allgather, MPI_Allgatherv, to gather data from
different processes into a single structure that is then sent
to all participants (Gather-to-all).
 MPI_Alltoall, MPI_Alltoallv, to gather data and
then scatter it to all participants (All-to-all
scatter/gather).
MPI_Bcast, to Broadcast data to other processes.
COMMUNICATION
PATTERNS
Cluster Computer s working on four communication
patterns-
1. Single Direction Communication
2.Pair-based Communication
3.Pre-posted Communication
4.All-start Communication
SINGLE DIRECTION
COMMUNICATION
Processes are paired off, with the lower rank sending
message to the higher rank in a tight loop.
The individual pair synchronize before communication
begins.
PAIR-BASED
COMMUNICATION
Each process communicates with a small number of
remote processes in each communication phase.
Communication is paired, so that a given process is
both sending and receiving messages with exactly one
other process at a time, rotating to a new process when
communication is complete.
PRE-POSTED
Excepted message reception in the next communication
phase is computed before starting the computation
phase.
This guarantees that receive buffer will be available
during the communication phase.
ALL-START
COMMUNICATION
It is very much same as that of the pre-posted
communication but it does not guarantee that all
receives are pre-posted.
After the computation, MPI_WaitALL is called.
A call to MPI_WaitALL can be used to wait for all
pending operation in a list.
WORKING STRATEGY
 Installation of Ubuntu 10.04 LTS.
Installation of C in Ubuntu 10.04 LTS.
Use of terminal.
Installation of MPI_ch package on our Linux system.
Study of basic Linux command & other Linux features
Study of MPI, its basic commands & syntax.
Execution of basic Linux & MPI commands.
 Execution of matrix program using C on linux platform.
CONTINUE...
 Execution of basic programs using MPI.
 Execution of parallel computing.
 We will generate fault, then detect & at last, recover
them by assigning the task of faulty process to some
other process so as to overcome from failure.
 We will apply fault tolerance techniques i.e.
 Co-ordinate checkpoints
 Message logging
RESEARCH GAP
Up to now, fault tolerance has not yet been applied to
communication patterns.
So as to overcome with this problem, we need to introduce
fault tolerance in communication patterns so as to reach
to the correct final outcome.
GRAPHICAL
REPRESENTATION
0
50
100
150
200
250
300
350
400
may
apr
mar
feb
jan
dec
nov
oct
sept
Rgk cluster computing project

More Related Content

PDF
Inter process communication
PPT
Interprocess communication
PDF
MPI History
PPTX
Inter Process Communication
PPT
PPT
Chapter 6 pc
PPTX
Inter process communication
Interprocess communication
MPI History
Inter Process Communication
Chapter 6 pc

What's hot (17)

PPTX
Inter Process Communication-R.D.Sivakumar
PPT
Point-to-Point Communicationsin MPI
PPTX
Multiprocessing -Interprocessing communication and process sunchronization,se...
PDF
1844 1849
PPT
Interprocess communication (IPC) IN O.S
PDF
Client-centric Consistency Models
PPT
Consistency protocols
PDF
Replication in the Wild
PPTX
øMQ Vortrag
PDF
T3UC_PresentationT3UC2012_Rajesh
PPT
Group Communication (Distributed computing)
PDF
Distributed Systems Theory for Mere Mortals
PDF
MPI in TNT for parallel processing
DOCX
Process scheduling
DOCX
MC0085 – Advanced Operating Systems - Master of Computer Science - MCA - SMU DE
PPT
Client Centric Consistency Model
DOCX
Process synchronization
Inter Process Communication-R.D.Sivakumar
Point-to-Point Communicationsin MPI
Multiprocessing -Interprocessing communication and process sunchronization,se...
1844 1849
Interprocess communication (IPC) IN O.S
Client-centric Consistency Models
Consistency protocols
Replication in the Wild
øMQ Vortrag
T3UC_PresentationT3UC2012_Rajesh
Group Communication (Distributed computing)
Distributed Systems Theory for Mere Mortals
MPI in TNT for parallel processing
Process scheduling
MC0085 – Advanced Operating Systems - Master of Computer Science - MCA - SMU DE
Client Centric Consistency Model
Process synchronization
Ad

Viewers also liked (18)

PDF
HPC Cluster & Cloud Computing
PPTX
Cluster computing
PDF
Cluster Computing Environment for On - line Static Security Assessment of lar...
PPTX
Clustering by AKASHMSHAH
PDF
Cluster computing report
PPT
Cluster computing2
PPTX
PPTX
Cluster computing
PPT
Cluster Computing Seminar.
PPTX
Cluster computer
PPTX
Cluster computing
PPTX
Cluster computing
PPTX
Cluster computing
PPTX
cluster computing
PPT
Grid and cluster_computing_chapter1
PPTX
Cluster and Grid Computing
PPT
CLUSTER COMPUTING
PPT
Cluster Computing
HPC Cluster & Cloud Computing
Cluster computing
Cluster Computing Environment for On - line Static Security Assessment of lar...
Clustering by AKASHMSHAH
Cluster computing report
Cluster computing2
Cluster computing
Cluster Computing Seminar.
Cluster computer
Cluster computing
Cluster computing
Cluster computing
cluster computing
Grid and cluster_computing_chapter1
Cluster and Grid Computing
CLUSTER COMPUTING
Cluster Computing
Ad

Similar to Rgk cluster computing project (20)

PDF
Clustering Manual for Parallel Computing
PPTX
Introduction to MPI Basics easy way.pptx
PDF
Cluster Setup Manual Using Ubuntu and MPICH
PPTX
Clustering
PDF
Complete MPICH2 Clustering Manual in Ubuntu
PDF
MPI Presentation
PPTX
Presentation - Programming a Heterogeneous Computing Cluster
DOCX
Clustering manual
PDF
Move Message Passing Interface Applications to the Next Level
PPT
Open MPI
PPT
Parallel computing(2)
PPTX
Distributed Memory Programming with MPI
PDF
Using MPI third edition Portable Parallel Programming with the Message Passin...
PDF
Building A Linux Cluster Using Raspberry PI #2!
PPT
chapter4.ppt
PPT
59137949-Cluster-Computing (1).ppt .....
PPTX
25-MPI-OpenMP.pptx
PDF
Wondershare Filmora Crack 2025 For Windows Free
PPTX
Smalland Survive the Wilds v1.6.2 Free Download
Clustering Manual for Parallel Computing
Introduction to MPI Basics easy way.pptx
Cluster Setup Manual Using Ubuntu and MPICH
Clustering
Complete MPICH2 Clustering Manual in Ubuntu
MPI Presentation
Presentation - Programming a Heterogeneous Computing Cluster
Clustering manual
Move Message Passing Interface Applications to the Next Level
Open MPI
Parallel computing(2)
Distributed Memory Programming with MPI
Using MPI third edition Portable Parallel Programming with the Message Passin...
Building A Linux Cluster Using Raspberry PI #2!
chapter4.ppt
59137949-Cluster-Computing (1).ppt .....
25-MPI-OpenMP.pptx
Wondershare Filmora Crack 2025 For Windows Free
Smalland Survive the Wilds v1.6.2 Free Download

Recently uploaded (20)

PDF
BIO-INSPIRED ARCHITECTURE FOR PARSIMONIOUS CONVERSATIONAL INTELLIGENCE : THE ...
PPTX
Fundamentals of Mechanical Engineering.pptx
PDF
PREDICTION OF DIABETES FROM ELECTRONIC HEALTH RECORDS
PPT
introduction to datamining and warehousing
PDF
SMART SIGNAL TIMING FOR URBAN INTERSECTIONS USING REAL-TIME VEHICLE DETECTI...
PDF
III.4.1.2_The_Space_Environment.p pdffdf
PDF
Soil Improvement Techniques Note - Rabbi
PPT
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
PPT
INTRODUCTION -Data Warehousing and Mining-M.Tech- VTU.ppt
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
Nature of X-rays, X- Ray Equipment, Fluoroscopy
PDF
PPT on Performance Review to get promotions
PDF
Categorization of Factors Affecting Classification Algorithms Selection
PDF
Abrasive, erosive and cavitation wear.pdf
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PPTX
UNIT 4 Total Quality Management .pptx
PPTX
introduction to high performance computing
PDF
Integrating Fractal Dimension and Time Series Analysis for Optimized Hyperspe...
PPTX
Safety Seminar civil to be ensured for safe working.
PPT
Total quality management ppt for engineering students
BIO-INSPIRED ARCHITECTURE FOR PARSIMONIOUS CONVERSATIONAL INTELLIGENCE : THE ...
Fundamentals of Mechanical Engineering.pptx
PREDICTION OF DIABETES FROM ELECTRONIC HEALTH RECORDS
introduction to datamining and warehousing
SMART SIGNAL TIMING FOR URBAN INTERSECTIONS USING REAL-TIME VEHICLE DETECTI...
III.4.1.2_The_Space_Environment.p pdffdf
Soil Improvement Techniques Note - Rabbi
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
INTRODUCTION -Data Warehousing and Mining-M.Tech- VTU.ppt
R24 SURVEYING LAB MANUAL for civil enggi
Nature of X-rays, X- Ray Equipment, Fluoroscopy
PPT on Performance Review to get promotions
Categorization of Factors Affecting Classification Algorithms Selection
Abrasive, erosive and cavitation wear.pdf
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
UNIT 4 Total Quality Management .pptx
introduction to high performance computing
Integrating Fractal Dimension and Time Series Analysis for Optimized Hyperspe...
Safety Seminar civil to be ensured for safe working.
Total quality management ppt for engineering students

Rgk cluster computing project

  • 2. FAULT TOLERANCE IN CLUSTER COMPUTING Guided By- Submitted By- Mr. Ankush Agrawal Ravindra Pratap Singh Mr. Praveen Rai Garima Kaushik Kamini Saraswat
  • 3. OUTLINE  Introduction Purpose Requirements Advantages Of Linux  Objective  Sub-Objective  Research Gap  Basic MPI Commands Message Passing Interface Working Strategy Graphical Representation
  • 4. INTRODUCTION What Is Cluster …??? A cluster is a set of connected computers that work together so that it can be viewed as a single system. It works on master slave connection.  What Is Cluster Computing…??? A cluster computing is also known as HPC as it is used to solve the large problems in less time compared with other techniques. HPC may include Parallel, Cluster, Grid, Cloud and Green computing.
  • 5. CONTINUE... What Is Fault…??? A fault is any error or unwanted condition that may arise in a system due to which our system will stop its execution. It may be a natural or man-made types.  What Is Fault Tolerance…??? A fault tolerance is an ability by which we will tolerate some type of faults so that we will get the correct final outcome. Eg. Faulty processor etc.
  • 6. PUPOSE The purpose of cluster technology is to eliminate single points of failure. When availability of data is your paramount consideration, clustering is ideal. Using a cluster we can avoids all of these single points of failure:  Network card failure  Processor failure  Motherboard failure
  • 7. REQUIREMENTS Software Environment Operating system- Ubuntu 10.0.4 LTS MPI_Ch2 Package Open_MPI Libshem-dev Libmpich2-dev
  • 8. ADVANTAGES OF USING LINUX The following are some advantages of using Linux: Linux is readily available on the Internet and can be downloaded without cost. It is easy to fix bugs and improve system performance. Users can develop or fine-tune hardware drivers which can easily be made available to other users. The most important advantage of using linux is that it creates a several copies of one processor which helps in enhancing the performance of a system.
  • 9. OBJECTIVE We are working on linux operating system & on a communication patterns of clusters using MPI. Our aim is to find faults, and to recover those faults which are causing unexpected behaviours (error , bugs etc.).
  • 10. MESSAGE PASSING INTERFACE(MPI) The generic form of message passing in the parallel computing is the Message Passing Interface. It is used as a medium of communication among the nodes. In message passing, data is moved from address space of one to that of other by mean of cooperative operation such as send/receive pair.
  • 11. BASIC MPI ROUTINS/COMMANDS For comunication among different processes some routines are used which are- MPI_Send, to send a message to another process. MPI_Recv, to receive a message from another process. MPI_Gather, MPI_Gatherv, to gather data from participating processes into a single structure.  MPI Comm size() – Number of MPI processes.  MPI Comm rank() – Internal process number.  MPI Get processor name() – External processor name.
  • 12. CONTINUE…  MPI_Scatter, MPI_Scatter, to break a structure into portions and distribute those portions to other processes.  MPI_Allgather, MPI_Allgatherv, to gather data from different processes into a single structure that is then sent to all participants (Gather-to-all).  MPI_Alltoall, MPI_Alltoallv, to gather data and then scatter it to all participants (All-to-all scatter/gather). MPI_Bcast, to Broadcast data to other processes.
  • 13. COMMUNICATION PATTERNS Cluster Computer s working on four communication patterns- 1. Single Direction Communication 2.Pair-based Communication 3.Pre-posted Communication 4.All-start Communication
  • 14. SINGLE DIRECTION COMMUNICATION Processes are paired off, with the lower rank sending message to the higher rank in a tight loop. The individual pair synchronize before communication begins.
  • 15. PAIR-BASED COMMUNICATION Each process communicates with a small number of remote processes in each communication phase. Communication is paired, so that a given process is both sending and receiving messages with exactly one other process at a time, rotating to a new process when communication is complete.
  • 16. PRE-POSTED Excepted message reception in the next communication phase is computed before starting the computation phase. This guarantees that receive buffer will be available during the communication phase.
  • 17. ALL-START COMMUNICATION It is very much same as that of the pre-posted communication but it does not guarantee that all receives are pre-posted. After the computation, MPI_WaitALL is called. A call to MPI_WaitALL can be used to wait for all pending operation in a list.
  • 18. WORKING STRATEGY  Installation of Ubuntu 10.04 LTS. Installation of C in Ubuntu 10.04 LTS. Use of terminal. Installation of MPI_ch package on our Linux system. Study of basic Linux command & other Linux features Study of MPI, its basic commands & syntax. Execution of basic Linux & MPI commands.  Execution of matrix program using C on linux platform.
  • 19. CONTINUE...  Execution of basic programs using MPI.  Execution of parallel computing.  We will generate fault, then detect & at last, recover them by assigning the task of faulty process to some other process so as to overcome from failure.  We will apply fault tolerance techniques i.e.  Co-ordinate checkpoints  Message logging
  • 20. RESEARCH GAP Up to now, fault tolerance has not yet been applied to communication patterns. So as to overcome with this problem, we need to introduce fault tolerance in communication patterns so as to reach to the correct final outcome.