SlideShare a Scribd company logo
1
Basic Components of a Parallel (or Serial) Computer
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
CPU
MEM
•Processors
• Memory
•Interconnect Network
2
Processor Related Terms
• RISC: Reduced Instruction Set Computers
• PIPELINE : Technique where multiple instructions
are overlapped in execution
• SUPERSCALAR: Multiple instructions per clock
period
3
Network Interconnect Related Terms
• LATENCY : How long does it take to start sending
a "message"? Units are generally microseconds
or milliseconds.
• BANDWIDTH : What data rate can be sustained
once the message is started? Units are bytes/sec,
Mbytes/sec, Gbytes/sec etc.
• TOPLOGY: What is the actual ‘shape’ of the
interconnect? Are the nodes connect by a 2D
mesh? A ring? Something more elaborate?
4
Memory/Cache Related Terms
CPU
MAIN MEMORY
Cache
CACHE : Cache is the level of memory hierarchy between
the CPU and main memory. Cache is much smaller than
main memory and hence there is mapping of data from
main memory to cache.
5
Memory/Cache Related Terms
• ICACHE : Instruction cache
• DCACHE (L1) : Data cache closest to registers
• SCACHE (L2) : Secondary data cache
– Data from SCACHE has to go through DCACHE
to registers
– SCACHE is larger than DCACHE
– All processors do not have SCACHE
• TLB : Translation-lookaside buffer keeps
addresses of pages ( block of memory) in main
memory that have recently been accessed
6
Memory/Cache Related Terms (cont.)
CPU
MEMORY
(e.g., L1 cache)
MEMORY
(e.g., L2 cache)
MEMORY
(e.g., DRAM)
SPEED SIZE Cost ($/bit)
7
Memory/Cache Related Terms (cont.)
• The data cache was designed with two key concepts in
mind
– Spatial Locality
• When an element is referenced its neighbors will be
referenced too
• Cache lines are fetched together
• Work on consecutive data elements in the same cache
line
– Temporal Locality
• When an element is referenced, it might be referenced
again soon
• Arrange code so that date in cache is reused as often as
possible
8
Memory/Cache Related Terms (cont.)
Direct mapped cache: A block from main memory can go in
exactly one place in the cache. This is called direct mapped
because there is direct mapping from any block address in
memory to a single location in the cache.
cache
Main memory
9
Memory/Cache Related Terms (cont.)
Fully associative cache : A block from main memory can be
placed in any location in the cache. This is called fully
associative because a block in main memory may be
associated with any entry in the cache.
cache
Main memory
Memory/Cache Related Terms (cont.)
Set associative cache : The middle range of designs between
direct mapped cache and fully associative cache is called
set-associative cache. In a n-way set-associative cache a
block from main memory can go into n (n at least 2)
locations in the cache.
2-way set-associative cache
Main memory
Memory/Cache Related Terms (cont.)
• Least Recently Used (LRU) : Cache replacement
strategy for set associative caches. The cache
block that is least recently used is replaced with a
new block.
• Random Replace : Cache replacement strategy
for set associative caches. A cache block is
randomly replaced.
Types of Parallel Computers
• Until recently, Flynn's taxonomy was commonly
use to classify parallel computers into one of four
basic types:
– Single instruction, single data (SISD): single scalar
processor
– Single instruction, multiple data (SIMD): Thinking
machines CM-2
– Multiple instruction, single data (MISD): various
special purpose machines
– Multiple instruction, multiple data (MIMD): Nearly
all parallel machines
• However, since the MIMD model “won,” a much
more useful way to classify modern parallel
computers is by their memory model
– shared memory
– distributed memory
– (more recently) hybrid of the above two (also
called multi-tiered, CLUMPS )
P P P P P P
B U S / Cross Bar
M e m o r y
Shared and Distributed memory
Shared memory: single address
space. All processors have access
to a pool of shared memory.
(examples: Sun ES10000)
Methods of memory access :
- Bus
- Crossbar
Distributed memory: each processor
has it’s own local memory. Must do
message passing to exchange data
between processors.
(examples: IBM SP, CRAY T3E)
M
P
M
P
M
P
M
P
M
P
M
P
Network

More Related Content

PPT
Cache coherence
PPTX
Elements of cache design
PPTX
Cache coherence ppt
PPTX
Cache memory
PDF
Cache coherence
PPTX
Cache coherence
PPTX
Cache design
PPTX
Directory based cache coherence
Cache coherence
Elements of cache design
Cache coherence ppt
Cache memory
Cache coherence
Cache coherence
Cache design
Directory based cache coherence

What's hot (20)

PDF
Coherence and consistency models in multiprocessor architecture
PPTX
Address mapping
PPTX
Cache memory and virtual memory
PPTX
Cache Memory
PPT
Snooping 2
PDF
Lecture 6.1
PPT
Snooping protocols 3
PDF
Shared-Memory Multiprocessors
PPTX
Cache memoy designed by Mohd Tariq
PPTX
Cache memory principles
PPTX
Shared Memory Multi Processor
PPT
Cache memory and cache
PPTX
Cache memory
PPT
cache memory
PPT
cache memory management
PPTX
Cache Memory
PPTX
Cache memory
PPTX
Cachememory
PPTX
Coherence and consistency models in multiprocessor architecture
Address mapping
Cache memory and virtual memory
Cache Memory
Snooping 2
Lecture 6.1
Snooping protocols 3
Shared-Memory Multiprocessors
Cache memoy designed by Mohd Tariq
Cache memory principles
Shared Memory Multi Processor
Cache memory and cache
Cache memory
cache memory
cache memory management
Cache Memory
Cache memory
Cachememory
Ad

Viewers also liked (20)

PPT
Towards a Multilingual Ontology for Ontology-driven Content Mining in Social ...
DOCX
Clustering manual
PPTX
A Identificação de Riscos Novos e Potencializados em Projetos de Tecnologia d...
PPT
Lecture1
PPTX
A Multidomain and Multilingual Conceptual Data Model for Online Reviews Repre...
DOCX
Matrix multiplication graph
PPTX
Phd Marcirio Chaves
PPTX
Revisita e Análise dos Métodos para Captura de Lições Aprendidas: Uma Contrib...
PDF
A Fine-Grained Analysis of User-Generated Content to Support Decision Making
PPT
Parallel computing(1)
DOCX
Observer pattern
PPTX
Job search_resume
PDF
Database management system chapter15
PPT
Database management system chapter12
PPTX
Job search_interview
PPT
Tutorial on Parallel Computing and Message Passing Model - C5
PPT
Tutorial on Parallel Computing and Message Passing Model - C4
PPT
Collective Communications in MPI
PPT
Chapter 1 pc
PPT
Tutorial on Parallel Computing and Message Passing Model - C2
Towards a Multilingual Ontology for Ontology-driven Content Mining in Social ...
Clustering manual
A Identificação de Riscos Novos e Potencializados em Projetos de Tecnologia d...
Lecture1
A Multidomain and Multilingual Conceptual Data Model for Online Reviews Repre...
Matrix multiplication graph
Phd Marcirio Chaves
Revisita e Análise dos Métodos para Captura de Lições Aprendidas: Uma Contrib...
A Fine-Grained Analysis of User-Generated Content to Support Decision Making
Parallel computing(1)
Observer pattern
Job search_resume
Database management system chapter15
Database management system chapter12
Job search_interview
Tutorial on Parallel Computing and Message Passing Model - C5
Tutorial on Parallel Computing and Message Passing Model - C4
Collective Communications in MPI
Chapter 1 pc
Tutorial on Parallel Computing and Message Passing Model - C2
Ad

Similar to Lecture2 (20)

PDF
Computer architecture
PPTX
Computer System Architecture Lecture Note 8.2 Cache Memory
PDF
Computer architecture for HNDIT
PPTX
UNIT IV Computer architecture Analysis.pptx
PPT
Memory organization including cache and RAM.ppt
PDF
Memory (Computer Organization)
PPTX
CS304PC:Computer Organization and Architecture Session 29 Memory organization...
PDF
22CS201 COA
PPTX
BCSE205L_Module 4 Computer Architecture Org.pptx
PPTX
Cache Memory.pptx
PPT
Memory Organization and Cache mapping.ppt
PPT
Akanskaha_ganesh_kullarni_memory_computer.ppt
PPT
chapter-6-multiprocessors-and-thread-level (1).ppt
PPT
Massively Parallel Architectures
PPTX
Cache.pptx
PPTX
Computer organizatin.Chapter Seven.pptxs
PPT
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
PPT
Introduction to symmetric multiprocessor
PPT
cache Computer architecture and Org.ppt
PDF
Parallel Computing - Lec 3
Computer architecture
Computer System Architecture Lecture Note 8.2 Cache Memory
Computer architecture for HNDIT
UNIT IV Computer architecture Analysis.pptx
Memory organization including cache and RAM.ppt
Memory (Computer Organization)
CS304PC:Computer Organization and Architecture Session 29 Memory organization...
22CS201 COA
BCSE205L_Module 4 Computer Architecture Org.pptx
Cache Memory.pptx
Memory Organization and Cache mapping.ppt
Akanskaha_ganesh_kullarni_memory_computer.ppt
chapter-6-multiprocessors-and-thread-level (1).ppt
Massively Parallel Architectures
Cache.pptx
Computer organizatin.Chapter Seven.pptxs
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
Introduction to symmetric multiprocessor
cache Computer architecture and Org.ppt
Parallel Computing - Lec 3

More from tt_aljobory (20)

PDF
Homework 2 sol
PPT
Lecture12
PPT
Lecture11
PPT
Lecture10
PPT
Lecture9
PPT
Lecture7
PPT
Lecture8
PPT
Lecture6
PPT
Lecture5
PPT
Lecture4
PPT
Lecture3
PDF
Lecture 1
PDF
Good example on ga
PPT
PPT
PPT
PDF
Above theclouds
PDF
Inet prog
PDF
PDF
8051 experiments1
Homework 2 sol
Lecture12
Lecture11
Lecture10
Lecture9
Lecture7
Lecture8
Lecture6
Lecture5
Lecture4
Lecture3
Lecture 1
Good example on ga
Above theclouds
Inet prog
8051 experiments1

Lecture2

  • 1. 1 Basic Components of a Parallel (or Serial) Computer CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM •Processors • Memory •Interconnect Network
  • 2. 2 Processor Related Terms • RISC: Reduced Instruction Set Computers • PIPELINE : Technique where multiple instructions are overlapped in execution • SUPERSCALAR: Multiple instructions per clock period
  • 3. 3 Network Interconnect Related Terms • LATENCY : How long does it take to start sending a "message"? Units are generally microseconds or milliseconds. • BANDWIDTH : What data rate can be sustained once the message is started? Units are bytes/sec, Mbytes/sec, Gbytes/sec etc. • TOPLOGY: What is the actual ‘shape’ of the interconnect? Are the nodes connect by a 2D mesh? A ring? Something more elaborate?
  • 4. 4 Memory/Cache Related Terms CPU MAIN MEMORY Cache CACHE : Cache is the level of memory hierarchy between the CPU and main memory. Cache is much smaller than main memory and hence there is mapping of data from main memory to cache.
  • 5. 5 Memory/Cache Related Terms • ICACHE : Instruction cache • DCACHE (L1) : Data cache closest to registers • SCACHE (L2) : Secondary data cache – Data from SCACHE has to go through DCACHE to registers – SCACHE is larger than DCACHE – All processors do not have SCACHE • TLB : Translation-lookaside buffer keeps addresses of pages ( block of memory) in main memory that have recently been accessed
  • 6. 6 Memory/Cache Related Terms (cont.) CPU MEMORY (e.g., L1 cache) MEMORY (e.g., L2 cache) MEMORY (e.g., DRAM) SPEED SIZE Cost ($/bit)
  • 7. 7 Memory/Cache Related Terms (cont.) • The data cache was designed with two key concepts in mind – Spatial Locality • When an element is referenced its neighbors will be referenced too • Cache lines are fetched together • Work on consecutive data elements in the same cache line – Temporal Locality • When an element is referenced, it might be referenced again soon • Arrange code so that date in cache is reused as often as possible
  • 8. 8 Memory/Cache Related Terms (cont.) Direct mapped cache: A block from main memory can go in exactly one place in the cache. This is called direct mapped because there is direct mapping from any block address in memory to a single location in the cache. cache Main memory
  • 9. 9 Memory/Cache Related Terms (cont.) Fully associative cache : A block from main memory can be placed in any location in the cache. This is called fully associative because a block in main memory may be associated with any entry in the cache. cache Main memory
  • 10. Memory/Cache Related Terms (cont.) Set associative cache : The middle range of designs between direct mapped cache and fully associative cache is called set-associative cache. In a n-way set-associative cache a block from main memory can go into n (n at least 2) locations in the cache. 2-way set-associative cache Main memory
  • 11. Memory/Cache Related Terms (cont.) • Least Recently Used (LRU) : Cache replacement strategy for set associative caches. The cache block that is least recently used is replaced with a new block. • Random Replace : Cache replacement strategy for set associative caches. A cache block is randomly replaced.
  • 12. Types of Parallel Computers • Until recently, Flynn's taxonomy was commonly use to classify parallel computers into one of four basic types: – Single instruction, single data (SISD): single scalar processor – Single instruction, multiple data (SIMD): Thinking machines CM-2 – Multiple instruction, single data (MISD): various special purpose machines – Multiple instruction, multiple data (MIMD): Nearly all parallel machines
  • 13. • However, since the MIMD model “won,” a much more useful way to classify modern parallel computers is by their memory model – shared memory – distributed memory – (more recently) hybrid of the above two (also called multi-tiered, CLUMPS )
  • 14. P P P P P P B U S / Cross Bar M e m o r y Shared and Distributed memory Shared memory: single address space. All processors have access to a pool of shared memory. (examples: Sun ES10000) Methods of memory access : - Bus - Crossbar Distributed memory: each processor has it’s own local memory. Must do message passing to exchange data between processors. (examples: IBM SP, CRAY T3E) M P M P M P M P M P M P Network