Grenoble:Network: Difference between revisions

From Grid5000
Jump to navigation Jump to search
No edit summary
No edit summary
Line 17: Line 17:
The topology used is a fat tree with a 2:1 blocking factor:
The topology used is a fat tree with a 2:1 blocking factor:
* 2 top switchs
* 2 top switchs
* 6 leaves: switchs with 32 downlinks to nodes and 8 uplinks to each top switch (48 ports total)
* 6 leaves: switchs with 32 downlinks to nodes and 8 uplinks to each of the two top switches (48 ports total)


All 32 dahu nodes are connected to a same leaf. The 4 yeti nodes are connected to another leaf. TBC for lama.
All 32 dahu nodes are connected to a same leaf. The 4 yeti nodes are connected to another leaf. TBC for lama.
Other ports are used by nodes of the HPC center of UGA.


[[File:g5k-grenoble-network-opa.png]]
[[File:g5k-grenoble-network-opa.png]]

Revision as of 15:37, 3 December 2019

Overview of Ethernet network topology

GrenobleNetwork.png


Network devices models

  • gw: Dell S5296F-ON

More details (including address ranges) are available from the Grid5000:Network page.

Omni-Path Topology

Nodes of the dahu, yeti and lama clusters are connected to Omni-Path switches (Intel Omni-Path 100Gbps), beside the Ethernet network.

This Omni-Path network interconnects both the Grid'5000 Grenoble nodes and the nodes of the HPC Center of Université Grenoble Alpes (Gricad mésocentre).

The topology used is a fat tree with a 2:1 blocking factor:

  • 2 top switchs
  • 6 leaves: switchs with 32 downlinks to nodes and 8 uplinks to each of the two top switches (48 ports total)

All 32 dahu nodes are connected to a same leaf. The 4 yeti nodes are connected to another leaf. TBC for lama.

Other ports are used by nodes of the HPC center of UGA.

G5k-grenoble-network-opa.png


A generated topology (using the output of opareports -o topology :

Grenoble-DC-IMAG-OPA.svg

IP Range

  • Computing : 172.16.16.0/20
  • Omni-Path : 172.18.16.0/20
  • Virtual : 10.132.0.0/14