Skip to content

22. Cubes

Mar 29, 2021

Measuring Performance

of a message passing system

There are 3 factors to consider:

  • (1) message delay = latency
  • (2) number of messages
  • (3) local computation

Time in a message passing system means = message delay + local computation

  • broadcast latency is Θ(N)
  • delays vary by algorithms (measured as d), e.g.
    • ABD performs 1 round communications (2 delays/round)
    • ERATO takes 1.5 rounds (3 delays/round)

img

Architectures Comparison

  • WPRAM=PTP
  • WSIM=PTPL where L is the simulation latency


3D Cube

  • overhead: P3
  • HW cost: HW$=Θ(P)+Θ(6P)=Θ(P)
  • Latency: L=3P33

img

d L HW$ Wires/P Wires/Total
1 1(P1) P 2 2P
2 2(P1) P 4 4P
3 3(P31) P 6 6P
d d(Pd1) P 2d 2dP

How small can Pd be?

When d then L0

Pd1=1

Pd=2

P=22

d=logP

Must be at least 1.

Hypercube

d P name
0 P=20=1 img empty string λ
1 P=21=2 img 0 ... 1
2 P=22=4 img 00 ... 11
3 P=23=8 img 000 ... 111

Any hypercube is a multidimensional grid; not all grids are hypercubes.

Hypercube is logarithmically efficient for simulation:

  • great in proctice
  • cost in scalability
  • number of wires is 2log2P
  • need to add ports to connect processors

Length of node name is |name|=log2P

Cube-Connected Cycles

img

Hypercube of P nodes

  • PlogP nodes and 3 wires/processor
  • number of wires is not dependent on P anymore.

This works in other dimensions for hypercubes e.g. 8 node hypercube becomes 24 nodes. Number of wires per processor is 3.

img

Routing is more complicated now, e.g. routing from 000 to 111

img

  • diameter of graph: max of any distance between two nodes
  • diameter of hypercube = d
  • diameter of CCC = O(d2)
  • |name|=log2P+log2log2P