Unit 5
Unit 5
UNIT-V
Syllabus
• Pipeline and Vector Processing: Parallel Processing, Pipelining, Arithmetic Pipeline, Instruction Pipeline, RISC Pipeline, Vector
Processing, Array Processor. Multi Processors: Characteristics of Multiprocessors, Interconnection Structures, Interprocessor
arbitration, Interprocessor communication and synchronization, Cache Coherence.
G.SWARNALATHA,Asst.Professor, GNITC
V
List of Topics
Reduced Instruction Set Computer:
•CISC Characteristics,
•RISC Characteristics.
Pipeline and Vector Processing:
•Parallel Processing
•Pipelining, Arithmetic Pipeline
•Instruction Pipeline
•RISC Pipeline
•Vector Processing
•Array Processor
Multi Processors
•Characteristics of Multiprocessors
•Interconnection Structures
• Interprocessor arbitration,
• Interprocessor communication and synchronization
• Cache Coherence. G.SWARNALATHA,Asst.Professor, GNITC
V RISC and CISC
G.SWARNALATHA,Asst.Professor, GNITC
V CISC
CISC Characteristics
• One reason for the trend to provide a complex instruction set is
the desire to simplify the compilation and improve the overall
computer performance.
• The task of a compiler is to generate a sequence of machine
instructions for each high-level language statement.
• The task is simplified if there are machine instructions that
implement the statements directly.
• The essential goal of a CISC architecture is to attempt to provide a
single machine instruction for each statement that is written in a
high-level language.
• Examples of CISC architectures are the Digital Equipment Corporation VAX computer and the
IBM 370 computer.
G.SWARNALATHA,Asst.Professor, GNITC
V CISC
CISC Characteristics
G.SWARNALATHA,Asst.Professor, GNITC
V CISC
CISC Characteristics
G.SWARNALATHA,Asst.Professor, GNITC
V RISC
RISC Characteristics
• The concept of RISC architecture involves an attempt to reduce execution time
by simplifying the instruction set of the computer.
The major characteristics of a RISC processor are:
1. Relatively few instructions
2. Relatively few addressing modes
3. Memory access limited to load and store instructions
4. All operations done within the registers of the CPU
5. Fixed-length, easily decoded instruction format
6. Single-cycle instruction execution
7. Hardwired rather than microprogrammed control
8. A relatively large number of registers in the processor unit
9. Efficient instruction pipeline
G.SWARNALATHA,Asst.Professor, GNITC
V RISC
RISC Characteristics
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing
V
PIPELINING AND VECTOR PROCESSING
• Parallel Processing
• Pipelining
• Arithmetic Pipeline
• Instruction Pipeline
• RISC Pipeline
• Vector Processing
• Array Processors
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
PARALLEL PROCESSING
- Inter-Instruction level
- Intra-Instruction level
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
PARALLEL COMPUTERS
Architectural Classification
• Flynn's classification
• Based on the multiplicity of Instruction Streams and Data Streams
• Instruction Stream
• Sequence of Instructions read from memory
• Data Stream
• Operations performed on the data in the processor
VLIW
MISD Nonexistence
Systolic arrays
Dataflow
Associative processors
Message-passing multicomputers
Hypercube
Mesh
Reconfigurable
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
SISD COMPUTER SYSTEMS
Control Processor Data stream Memory
Unit Unit
Instruction stream
Characteristics
Limitations
• Multiprogramming
• Spooling
• Multifunction processor
• Pipelining
• Exploiting instruction-level parallelism
- Superscalar
- Superpipelining
- VLIW (Very Long Instruction Word)
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
M CU P
M CU P
Memory
• •
• •
• •
M CU Data stream
P
Instruction stream
Characteristics
- There is no computer at present that can be
classified as MISD
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
SIMD COMPUTER SYSTEMS
Memory
Data bus
Control Unit
Instruction stream
Data stream
Alignment network
Characteristics
- Only one copy of the program exists
- A single controller executes one instruction at a time
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
TYPES OF SIMD COMPUTERS
Array Processors
Systolic Arrays
Associative Processors
- Content addressing
- Data transformation operations over many sets
of arguments with a single instruction G.SWARNALATHA,Asst.Professor, GNITC
- STARAN, PEPE
Pipelining and Vector Processing Parallel Processing
V
MIMD COMPUTER SYSTEMS
P M P M ••• P M
Interconnection Network
Shared Memory
Characteristics
- Multiple processing units
- Message-passing multicomputers
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
SHARED MEMORY MULTIPROCESSORS
M M ••• M
Buses,
Interconnection Network(IN) Multistage IN,
Crossbar Switch
P P ••• P
Characteristics
All processors have equally direct access to
one large memory address space
Example systems
Bus and cache-based systems
- Sequent Balance, Encore Multimax
Multistage IN-based systems
- Ultracomputer, Butterfly, RP3, HEP
Crossbar switch-based systems
- C.mmp, Alliant FX/8
Limitations
Memory access latency
Hot spot problem
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Parallel Processing
V
MESSAGE-PASSING MULTICOMPUTER
Message-Passing Network Point-to-point connections
P P ••• P
M M ••• M
Characteristics
- Interconnected computers
- Each processor has its own memory, and
communicate via message-passing
Example systems
- Tree structure: Teradata, DADO
- Mesh-connected: Rediflow, Series 2010, J-Machine
- Hypercube: Cosmic Cube, iPSC, NCUBE, FPS T Series, Mark III
Limitations
- Communication overhead
- Hard to programming G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Pipelining
V
PIPELINING
A technique of decomposing a sequential process
into suboperations, with each subprocess being
executed in a partial dedicated segment that
operates concurrently with all other segments.
Ai * Bi + Ci for i = 1, 2, 3, ... , 7
Ai Bi Memory Ci
Segment 1
R1 R2
Multiplier
Segment 2
R3 R4
Adder
Segment 3
R5
Clock
Segment 1 Segment 2 Segment 3
Pulse
Number R1 R2 R3 R4 R5
1 A1 B1
2 A2 B2 A1 * B1 C1
3 A3 B3 A2 * B2 C2 A1 * B1 + C1
4 A4 B4 A3 * B3 C3 A2 * B2 + C2
5 A5 B5 A4 * B4 C4 A3 * B3 + C3
6 A6 B6 A5 * B5 C5 A4 * B4 + C4
7 A7 B7 A6 * B6 C6 A5 * B5 + C5
8 A7 * B7 C7 A6 * B6 + C6
9 A7 * B7 + C7
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Pipelining
V
GENERAL PIPELINE
General Structure of a 4-Segment Pipeline
Clock
Input S1 R1 S2 R2 S3 R3 S4 R4
Space-Time Diagram
1 2 3 4 5 6 7 8 9
Clock cycles
Segment 1 T1 T2 T3 T4 T5 T6
2 T1 T2 T3 T4 T5 T6
3 T1 T2 T3 T4 T5 T6
4 T1 T2 T3 T4 T5 T6
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Pipelining
V
PIPELINE SPEEDUP
n: Number of tasks to be performed
Speedup
Sk: Speedup
Sk = n*tn / (k + n - 1)*tp
tn
lim Sk = ( = k, if tn = k * tp )
n tp
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Pipelining
V
PIPELINE AND MULTIPLE FUNCTION UNITS
Example
- 4-stage pipeline
- subopertion in each stage; tp = 20nS
- 100 tasks to be executed
- 1 task in non-pipelined system; 20*4 = 80nS
Pipelined System
(k + n - 1)*tp = (4 + 99) * 20 = 2060nS
Non-Pipelined System
n*k*tp = 100 * 80 = 8000nS
Speedup
Sk = 8000 / 2060 = 3.88
Ii I i+1 I i+2 I i+3
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Arithmetic Pipeline
V
ARITHMETIC PIPELINE
Floating-point adder Exponents Mantissas
a b A B
X = A x 2a
Y = B x 2b
R R
R R
R R
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Arithmetic Pipeline
V
4-STAGE FLOATING POINT ADDER
A = a x 2p B = b x 2q
p a q b
Stages: Other
Exponent fraction Fraction
S1 subtractor selector
Fraction with min(p,q)
r = max(p,q)
Right shifter
t = |p - q|
S2 Fraction
adder
r c
Leading zero
S3 counter
c
Left shifter
r
d
Exponent
S4 adder
s d
C = A + B = c x 2r = d x 2s
(r = max (p,q), 0.5 d < 1) G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Instruction Pipeline
V
INSTRUCTION CYCLE
Six Phases* in an Instruction Cycle
[1] Fetch an instruction from memory
[2] Decode the instruction
[3] Calculate the effective address of the operand
[4] Fetch the operands from memory
[5] Execute the operation
[6] Store the result in the proper place
i FI DA FO EX
i+1 FI DA FO EX
i+2 FI DA FO EX
Pipelined
i FI DA FO EX
i+1 FI DA FO EX
i+2 FI DA FO EX
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Instruction Pipeline
V
INSTRUCTION EXECUTION IN A 4-STAGE PIPELINE
Decode instruction
Segment2: and calculate
effective address
yes Branch?
no
Fetch operand
Segment3: from memory
Interrupt yes
Interrupt?
handling
no
Update PC
Empty pipe
Step: 1 2 3 4 5 6 7 8 9 10 11 12 13
Instruction 1 FI DA FO EX
2 FI DA FO EX
(Branch) 3 FI DA FO EX
4 FI FI DA FO EX
5 FI DA FO EX
6 FI DA FO EX
G.SWARNALATHA,Asst.Professor, GNITC7 FI DA FO EX
Pipelining and Vector Processing Instruction Pipeline
V
MAJOR HAZARDS IN PIPELINED EXECUTION
Structural hazards(Resource Conflicts)
Hardware Resources required by the instructions in
simultaneous overlapped execution cannot be met
Data hazards (Data Dependency Conflicts)
An instruction scheduled to be executed in the pipeline requires the
result of a previous instruction, which is not yet available
bubble IF ID OF OE OS
i+1 FI DA FO EX
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Instruction Pipeline
V
DATA HAZARDS
Data Hazards
Interlock
- hardware detects the data dependencies and delays the scheduling
of the dependent instruction by stalling enough clock cycles
Forwarding (bypassing, short-circuiting)
- Accomplished by a data path that routes a value from a source
(usually an ALU) to a user, bypassing a designated register. This
allows the value to be produced to be used at an earlier stage in the
pipeline than would otherwise be possible
Example: Register
file
ADD R1, R2, R3
SUB R4, R1, R5
ALU Operations
E: Write the result to the
R4
destination register
ALU result buffer
ADD I A E
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Instruction Pipeline
V
INSTRUCTION SCHEDULING
a = b + c;
d = e - f;
Delayed Load
A load requiring that the following instruction not use its result
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Instruction Pipeline
V
CONTROL HAZARDS
Branch Instructions
Next
FI DA FO EX
Instruction
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing RISC Pipeline
V
RISC PIPELINE
RISC
- Machine with a very fast clock cycle that
executes at the rate of one instruction per cycle
<- Simple Instruction Set
Fixed Length Instruction Format
Register-to-Register Operations
Instruction Cycles of Three-Stage Instruction Pipeline
Data Manipulation Instructions
I: Instruction Fetch
A: Decode, Read Registers, ALU Operations
E: Write a Register
clock cycle 1 2 3 4 5 6
Load R1 I A E
Load R2 I A E
Add R1+R2 I A E
Store R3 I A E
VECTOR PROCESSING
Vector Processing Applications
• Problems that can be efficiently formulated in terms of vectors
• Long-range weather forecasting
• Petroleum explorations
• Seismic data analysis
• Medical diagnosis
• Aerodynamics and space flight simulations
• Artificial intelligence and expert systems
• Mapping the human genome
• Image processing
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Vector Processing
V
VECTOR PROGRAMMING
DO 20 I = 1, 100
20 C(I) = B(I) + A(I)
Conventional computer
Initialize I = 0
20 Read A(I)
Read B(I)
Store C(I) = A(I) + B(I)
Increment I = i + 1
If I 100 goto 20
Vector computer
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Vector Processing
V
VECTOR INSTRUCTIONS
f1: V * V
f2: V * S
f3: V x V * V V: Vector operand
f4: V x S * V S: Scalar operand
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Vector Processing
V
VECTOR INSTRUCTION FORMAT
Source
A
G.SWARNALATHA,Asst.Professor, GNITC
Pipelining and Vector Processing Vector Processing
V MULTIPLE MEMORY MODULE AND INTERLEAVING
AR AR AR AR
DR DR DR DR
Data bus
Address Interleaving
• Characteristics of Multiprocessors
• Interconnection Structures
• Interprocessor Arbitration
• Interprocessor Communication
and Synchronization
• Cache Coherence
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Characteristics of Multiprocessors
V TERMINOLOGY
Parallel Computing
Distributed Computing
Concurrent Computing
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Characteristics of Multiprocessors
V TERMINOLOGY
Supercomputing
Use of fastest, biggest machines to solve big, computationally
intensive problems. Historically machines were vector computers,
but parallel/vector or parallel becoming the norm
Pipelining
Breaking a task into steps performed by different units, and multiple
inputs stream through the units, with next input starting in a unit when
previous input done with the unit but not necessarily done with the task
Vector Computing
Use of vector processors, where operation such as multiply
broken into several steps, and is applied to a stream of operands
(“vectors”). Most common special case of pipelining
Systolic
Similar to pipelining, but units are not necessarily arranged linearly,
steps are typically small and more numerous, performed in lockstep
fashion. Often used in special-purpose hardware such as image or signal
processors
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Characteristics of Multiprocessors
V
SPEEDUP AND EFFICIENCY
A: Given problem
1 2 3 4 5 6 7 8 9 10
Processors
Speedup should be between 0 and p, and
Efficiency should be between 0 and 1
Speedup is linear if there is a constant c > 0
so that speedup is always at least cp. G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Characteristics of Multiprocessors
V AMDAHL’S LAW
Given a program
f : Fraction of time that represents operations
that must be performed serially
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Characteristics of Multiprocessors
V
GRANULARITY OF PARALLELISM
Granularity of Parallelism
Coarse-grain
Medium-grain
Fine-grain
SHARED MEMORY
DISTRIBUTED MEMORY
Memory
Network
Network
Processors G.SWARNALATHA,Asst.Professor,
Processors/Memory GNITC
Multiprocessors Characteristics of Multiprocessors
V
SHARED MEMORY MULTIPROCESSORS
M M M
...
Buses,
Interconnection Network Multistage IN,
Crossbar Switch
P P ... P
Characteristics
P P ... P
M M ... M
Characteristics
- Interconnected computers
- Each processor has its own memory, and
communicate via message-passing
Example systems
Limitations
Bus
All processors (and memory) are connected to a
common bus or busses
- Memory access is fairly uniform, but not very scalable
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V
BUS
- A collection of signal lines that carry module-to-module communication
- Data highways connecting several digital system elements
Operations of Bus
Devices
M3 S7 M6 S5 M4
S2
Bus
Local Bus
SYSTEM BUS
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V MULTIPORT MEMORY
Advantages
- Multiple paths -> high transfer rate Memory Modules
MM 1 MM 2 MM 3 MM 4
Disadvantages
- Memory control logic
- Large number of cables and
CPU 1
connections
CPU 2
CPU 3
CPU 4
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V
CROSSBAR SWITCH
Memory modules
CPU1
CPU2
CPU3
CPU4
} data,address, and
control from CPU 1
data
Memory
address
Multiplexers
and } data,address, and
control from CPU 2
Module R/W arbitration
logic
memory
enable
} data,address, and
control from CPU 3
} data,address, and
controlG.SWARNALATHA,Asst.Professor,
from CPU 4 GNITC
Multiprocessors Interconnection Structure
V MULTISTAGE SWITCHING NETWORK
Interstage Switch
0 0
A A
1 1
B B
A connected to 0 A connected to 1
0 0
A A
1 1
B B
B connected to 0 B connected to 1
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V MULTISTAGE INTERCONNECTION NETWORK
Binary Tree with 2 x 2 Switches 0
000
0 1
001
1
0
010
0
P1 1
1 011
P2
0
100
0
1
1 101
0
110
1
111
8x8 Omega Switching Network
0 000
1 001
2 010
3 011
4 100
5 101
6 110
7 111
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V HYPERCUBE INTERCONNECTION
- p = 2n
- processors are conceptually on the corners of a
n-dimensional hypercube, and each is directly
connected to the n neighboring nodes
- Degree = n
011 111
010
0 01 11 110
101
001
1 00 10 100
000
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interprocessor Arbitration
V
INTERPROCESSOR ARBITRATION
Bus
Board level bus
Backplane level bus
Interface level bus
Synchronous Bus
Each data item is transferred over a time slice
known to both source and destination unit
- Common clock source
- Or separate clock and synchronization signal
is transmitted periodically to synchronize
the clocks in the system
Asynchronous Bus
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interprocessor Arbitration
V
BUS SIGNALS
Miscellaneous control
Master clock CCLK
System initialization INIT
Byte high enable BHEN
Memory inhibit (2 lines) INH1 - INH2
Bus lock LOCK
Bus arbitration
Bus request BREQ
Common bus request CBRQ
Bus busy BUSY
Bus clock BCLK
Bus priority in BPRN
Bus priority out BPRO
Power and ground (20 lines)
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interprocessor Arbitration
V
INTERPROCESSOR ARBITRATION STATIC ARBITRATION
4x2
Priority encoder
2x4
Decoder
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interprocessor Arbitration
V
INTERPROCESSOR ARBITRATION DYNAMIC ARBITRATION
Time Slice
Fixed length time slice is given sequentially to
each processor, round-robin fashion
Polling
Unit address polling - Bus controller advances
the address to identify the requesting unit
LRU
FIFO
Receiving
Processor
Interrupt
Shared Memory
Receiving
Sending Processor
Communication Area
Processor
Instruction Mark
Receiver(s) Receiving
Processor
Message
..
.
Receiving
G.SWARNALATHA,Asst.Professor, GNITC Processor
Multiprocessors Interprocessor Communication and Synchronization
V
INTERPROCESSOR SYNCHRONIZATION
Synchronization
Communication of control information between processors
- To enforce the correct sequence of processes
- To ensure mutually exclusive access to shared writable data
Hardware Implementation
Semaphore
- A binary variable
- 1: A processor is executing a critical section,
that not available to other processors
0: Available to any requesting processor
- Software controlled Flag that is stored in G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interprocessor Communication and Synchronization
V
SEMAPHORE
Testing and Setting the Semaphore
These are being done while locked, so that other processors cannot test
and set while current processor is being executing these instructions
Bus
X = 52 X = 52 X = 52 Caches
P1 P2 P3 Processors
X = 120 X = 52 X = 52 Caches
P1 P2 P3 Processors
Bus
X = 120 X = 52 X = 52 Caches
P1 P2 G.SWARNALATHA,Asst.Professor,
P3 Processors GNITC
Multiprocessors Cache Coherence
V MAINTAINING CACHE COHERENCY
Shared Cache
- Disallow private cache
- Access time delay
Software Approaches
* Read-Only Data are Cacheable
- Private Cache is for Read-Only data
- Shared Writable Data are not cacheable
- Compiler tags data as cacheable and noncacheable
- Degrade performance due to software overhead
Hardware Approaches
* Snoopy Cache Controller
- Cache Controllers monitor all the bus requests from CPUs and IOPs
- All caches attached to the bus monitor the write operations
- When a word in a cache is written, memory is also updated (write through)
- Local snoopy controllers in all other caches check their memory to determine if they have
a copy of that word; If they have, that location is marked invalid(future reference to
this location causes cache miss) G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Parallel Computing
V
PARALLEL COMPUTING
Grosche’s Law
Minsky’s Conjecture
Amdahl’s Law
Software Inertia
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V INTERCONNECTION NETWORKS
Multistage Interconnect
Switch Processor
Bus
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V INTERCONNECTION NETWORKS
Direct Connection
Interconnection Network
A graph G(V,E)
V: a set of processors (nodes)
E: a set of wires (edges)
Ring
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V
INTERCONNECTION NETWORKS
• 2-Mesh
m
...
m
m2 = p
...
- Degree = 4
- Diameter = 2(m - 1)
- In general, an n-dimensional mesh has
diameter = d ( p1/n - 1)
- Diameter can be halved by having wrap-around
connections (-> Torus)
- Ring is a 1-dimensional mesh with wrap-around
connection G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V
INTERCONNECTION NETWORK
Binary Tree
- Degree = 3
p+1
- Diameter = 2 log
2
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors Interconnection Structure
V MIN SPACE
MIN
Banyan network
=(unique path network) Multiple Path Network
Permutation/Sorting Network
(N!)
• Clos network [53]
• Benes network [62]
• Batcher sorting
network [68]
G.SWARNALATHA,Asst.Professor, GNITC
Multiprocessors
V
SOME CURRENT PARALLEL COMPUTERS
DM-SIMD
• AMT DAP
• Goodyear MPP
• Thinking Machines CM series
• MasPar MP1
• IBM GF11
SM-MIMD
• Alliant FX
• BBN Butterfly
• Encore Multimax
• Sequent Balance/Symmetry
• CRAY 2, X-MP, Y-MP
• IBM RP3
• U. Illinois CEDAR
DM-MIMD
• Intel iPSC series, Delta machine
• NCUBE series
• Meiko Computing Surface
• Carnegie-Mellon/ Intel iWarp
G.SWARNALATHA,Asst.Professor, GNITC