Grid5000:Home: Difference between revisions

From Grid5000
Jump to navigation Jump to search
No edit summary
No edit summary
 
(72 intermediate revisions by 11 users not shown)
Line 2: Line 2:
{|width="95%"
{|width="95%"
|- valign="top"
|- valign="top"
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|bgcolor="#888888" style="border:1px solid #cccccc;padding:2em;padding-top:1em;"|
[[Image:Logo.png|left]]
[[File:Slices-ri-white-color.png|260px|left]]
<br>
<b>Grid'5000 is a precursor infrastructure of [http://www.slices-ri.eu SLICES-RI], Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.</b>
''Grid'5000 is a scientific instrument supporting experiment-driven research in all areas of computer science, including high performance computing, distributed computing, networking and big data.'' <br>
<br/>
[[media:seminaire_intro.pdf|Download the latest general introduction]], or a [https://www.grid5000.fr/screencast/index.html screencast of recent webUI developments]
Content on this website is partly outdated. Technical information remains relevant.
|}
|}
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
== Latest updates from Grid'5000 users ==
* '''Publications'''
{{#publications:3||**}}
==Latest news==
=== Journée Aramis ===


Le 16/04 aura lieu la journée Aramis "Usages et services du cloud dans notre environnement enseignement supérieur et recherche" avec une présentation de Grid'5000 par Simon Delamare.  
{|width="95%"
|- valign="top"
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
[[Image:g5k-backbone.png|thumbnail|260px|right|Grid'5000]]
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.'''


Cette journée est webcastée par la cellule Webcast du CC-IN2P3. Vous pouvez donc suivre en direct les [http://webcast.in2p3.fr/live/pleiniere_aramis_2014 présentations]. Le programme de la journée est disponible [http://aramis.resinfo.org/wiki/doku.php?id=pleniaires:pleniere17avril2014 ici]
Key features:
* provides '''access to a large amount of resources''': 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
* '''designed to support Open Science and reproducible research''', with full traceability of infrastructure and software changes on the testbed
* '''a vibrant community''' of 500+ users supported by a solid technical team


=== [[Grid5000:UserCharter|Grid'5000 Charter]] revised ===
<br>
The Grid'5000 charter has been reworded and simplified so as to be more inclusive about possible uses. The rules to ensure availability during the day for any user and during the night for large resource usage have not been changed.
Read more about our [[Team|teams]], our [[Publications|publications]], and the [[Grid5000:UsagePolicy|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].
=== [[Grid5000:school2014|Grid'5000 school 2014]] announced ===
Organized in Lyon, from June 16th to June 19th 2014, this [[Grid5000:school2014|spring school]] will bring together, but is not limited to, Grid'5000's users, technical team and executive committee for 4 days of tutorials and talks focusing on best-practices and results. Presentations and practical sessions will cover both basic usage of the platform, for new users, or potential users of Grid'5000 and advanced and new usage of the platform, for current users. Deadline to submit presentation proposals is March 29th, 2014.  


=== Grid'5000 users win second prize at CCGRID 2013's SCALE challenge ===
<br>
Snooze based entry running on Grid'5000 entry wins [http://www.pds.ewi.tudelft.nl/ccgrid2013/awards/ 2nd prize] at [http://www.pds.ewi.tudelft.nl/ccgrid2013/ CCGrid 2013] [http://www.pds.ewi.tudelft.nl/ccgrid2013/calls/scale-challenge/ SCALE challenge]: well done Matthieu and Anne-Cécile for defending the entry titled ''Scalability of the Snooze Autonomic Cloud Management System'' by Eugen Feller, Christine Morin, Matthieu Simonin, Anne-Cécile Orgerie, and Yvon Jégou.
Published documents and presentations:
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]


=== Grid'5000 users finalists of the SCALE'2013 challenge ===
Older documents:
Two submissions (out of five) from Grid'5000 users took part in the final of the international SCALE'2013 challenge (held with CCGrid'2013):
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
* D. Balouek, A. Lèbre, F. Quesnel Flauncher and DVMS -- Deploying and Scheduling Thousands of Virtual Machines on Hundreds of Nodes Distributed Geographically
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]
* Eugen Feller, Christine Morin, Matthieu Simonin, Anne-Cécile Orgerie, and Yvon Jégou -- Scalability of the Snooze Autonomic Cloud Management System.


The first proposition presents the deployment and scheduling of thousands of virtual machines, conducted with the Flauncher and DVMS frameworks, across the Grid'5000 testbed.
<br>
The frameworks have been able to deploy and schedule up to 10000 VMs during the tests. This research has been conducted in the context of the INRIA Hemera initiative.
Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL [[Hemera|HEMERA]] (2010-2014).
 
|}
The second proposition presents the Snooze architecture and focus on the following aspects :
* System set up scalability and resources consumption,
* Self-healing capabilities.
The deployment of Snooze were distributed over several sites of the Grid'5000 testbed.
The framework has been able to start 11000 of system services, recover thousands of failures
and used to launch large hadoop/mapreduce experiments.
This research has been conducted in the context of the INRIA Snooze ADT.
=== Grid'5000 tutorial during ComPAS'2013 ===
The [http://compas2013.inrialpes.fr/ ComPAS'2013 conference] (replacing RenPar, SympA and CFSE), to be held in Grenoble between January 15th and 18th, will feature a Grid'5000 tutorial.
----
[[Grid5000:News|read more news]]


<br>
<br>
==Grid'5000 at a glance==
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
<!-- [[Image:site_map.png|thumbnail|128px|right|Grid'5000 sites]] -->
<br>
[[Image:renater5-g5k.jpg|thumbnail|200px|right|Grid'5000]]


== Random pick of publications ==
{{#publications:}}


* '''Grid'5000''' is a scientific instrument for the study of large scale parallel and distributed systems. It aims at providing a '''highly reconfigurable, controlable and monitorable experimental platform''' to its users. The initial aim (circa 2003) was to reach 5000 processors in the platform. It has been reframed at 5000 cores, and was reached during winter 2008-2009.
==Latest news==
* The infrastructure of Grid'5000 is geographically distributed on different sites hosting the instrument, initially 9 sites in France (10 since 2011). Porto Alegre, Brazil is now officially becoming the first site abroad.
<rss max=4 item-max-length="2000">https://www.grid5000.fr/rss/G5KNews.php</rss>
----
[[News|Read more news]]


===Sites:===
=== Grid'5000 sites===
{|width="75%" cellspacing="3"  
{|width="100%" cellspacing="3"  
|- valign="top"
|- valign="top"
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Bordeaux:Home|Bordeaux]]
* [[Grenoble:Home|Grenoble]]
* [[Grenoble:Home|Grenoble]]
* [[Lille:Home|Lille]]
* [[Lille:Home|Lille]]
* [[Luxembourg:Home|Luxembourg]]
* [[Luxembourg:Home|Luxembourg]]
* [[Louvain:Home|Louvain]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Lyon:Home|Lyon]]
* [[Lyon:Home|Lyon]]
* [[Nancy:Home|Nancy]]
* [[Nancy:Home|Nancy]]
* [[Nantes:Home|Nantes]] (soon)
* [[Nantes:Home|Nantes]]
* [[Reims:Home|Reims]]
* [[Rennes:Home|Rennes]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Rennes:Home|Rennes]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Strasbourg:Home|Strasbourg]]
* [[Toulouse:Home|Toulouse]]
* [[Toulouse:Home|Toulouse]]
|-
|-
|}
|}
[[Image:Software layers.png|thumbnail|271px|left|Grid'5000 allows experiments in all these software layers]]
* '''Grid'5000''' is a research effort developing a '''large scale nation wide infrastructure for large scale parallel and distributed computing research'''.
* '''19 [[Grid5000:Laboratories|laboratories]]''' are involved in France with the objective of providing the community a testbed allowing experiments in all the software layers between the network protocols up to the applications.
The current plans are to extend from the 9 initial sites each with 100 to a thousand PCs, connected by the [http://www.renater.fr RENATER] Education and Research Network to a bigger platform including a few sites outside France not necessarily connected through a dedicated network connection. Sites in Brazil and Luxembourg should join shortly, and Reims has now joined.
All sites in France are connected to [http://www.renater.fr RENATER] with a 10Gb/s link, except Reims, for the time linked through a 1Gb/s
This high collaborative research effort is funded by INRIA, CNRS, the Universities of all sites and some regional councils.
== ALADDIN-G5K : ensuring the development of '''Grid'5000''' ==
For the 2008-2012 period, Engineers ensuring the development and day to day support of the infrastructure are mostly provided by INRIA, under the ''ADT ALADDIN-G5K''  initiative.
==[[Hemera|HEMERA: Demonstrating ambitious up-scaling techniques on '''Grid'5000''']] ==
[[Hemera|Héméra]] is an INRIA Large Wingspan project, started in 2010, that aims at demonstrating ambitious up-scaling techniques for large scale distributed computing by carrying out several dimensioning experiments on the Grid’5000 infrastructure, at animating the scientific community around Grid’5000 and at enlarging the Grid’5000 community by helping newcomers to make use of Grid’5000.
== Initial Rationale==
'''The foundations of Grid'5000''' have emerged from a thorough analysis and numerous discussions about methodologies used for scientific research in the Grid domain. A report presents the [http://www-sop.inria.fr/aci/grid/public/Library/rapport-grid5000-V3.pdf rationale for Grid'5000].
In addition to theory, simulators and emulators, there is a strong need for '''large scale testbeds''' where real life experimental conditions hold. '''The size of Grid'5000''', in terms of number of sites and number of processors per site, was established according to the scale of the experiments and the number of researchers involved in the project.


== Current funding ==
== Current funding ==
As from June 2008, INRIA is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].
{|width="100%" cellspacing="3"
{|width="100%" cellspacing="3"
|-
|-
Line 114: Line 78:
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===CNRS===
===CNRS===
[[Image:CNRS-filaire-MonoBleu.gif|100px]]
[[Image:CNRS-filaire-Quadri.png|125px]]
|-
|-
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Universities===
===Universities===
University Joseph Fourier, Grenoble<br/>
IMT Atlantique<br/>
University of Rennes 1, Rennes<br/>
Université Grenoble Alpes, Grenoble INP<br/>
Université Rennes 1, Rennes<br/>
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
University Bordeaux 1, Bordeaux<br/>
Université Bordeaux 1, Bordeaux<br/>
University Lille 1, Lille<br/>
Université Lille 1, Lille<br/>
Ecole Normale Supérieure, Lyon<br/>
École Normale Supérieure, Lyon<br/>
Université de Reims Champagne-Ardenne, Reims<br/>
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Regional councils===
===Regional councils===
Aquitaine<br/>
Aquitaine<br/>
Auvergne-Rhône-Alpes<br/>
Bretagne<br/>
Bretagne<br/>
Champagne-Ardenne<br/>
Champagne-Ardenne<br/>
Provence Alpes Côte d'Azur<br/>
Provence Alpes Côte d'Azur<br/>
Nord Pas de Calais<br/>
Hauts de France<br/>
Lorraine<br/>
Lorraine<br/>
|}
|}

Latest revision as of 09:50, 10 June 2025

Slices-ri-white-color.png

Grid'5000 is a precursor infrastructure of SLICES-RI, Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.
Content on this website is partly outdated. Technical information remains relevant.

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.

Key features:

  • provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.


Published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2025-07-01 07:22): 1 current events, 7 planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2777 overall):

  • Anna Kravchenko, Sjoerd Jacob De Vries, Malika Smaïl-Tabbone, Isaure Chauvot de Beauchêne. HIPPO: HIstogram-based Pseudo-POtential for scoring protein-ssRNA fragment-based docking poses. BMC Bioinformatics, 2024, 10.1186/s12859-024-05733-6. hal-04234486v2 view on HAL pdf
  • Chih-Kai Huang. Scalability of public geo-distributed fog computing federations. Other cs.OH. Université de Rennes, 2024. English. NNT : 2024URENS055. tel-04910860v2 view on HAL pdf
  • Chaima Zoghlami. Enhancing V2X communication systems for cooperative perception : road users safety. Library and information sciences. Université Paul Sabatier - Toulouse III, 2023. English. NNT : 2023TOU30172. tel-04344780 view on HAL pdf
  • Gustavo Salazar-Gomez, Wenqian Liu, Manuel Alejandro Diaz-Zapata, David Sierra González, Christian Laugier. TLCFuse: Temporal Multi-Modality Fusion Towards Occlusion-Aware Semantic Segmentation. IV 2024 - 35th IEEE Intelligent Vehicles Symposium, Jun 2024, Jeju Island, South Korea. pp.2110-2116, 10.1109/IV55156.2024.10588460. hal-04717193 view on HAL pdf
  • Miguel Felipe Silva Vasconcelos. Strategies for operating and sizing low-carbon cloud data centers. Other cs.OH. Université Grenoble Alpes 2020-..; Universidade de São Paulo (Brésil), 2023. English. NNT : 2023GRALM093. tel-04678116 view on HAL pdf


Latest news

Rss.svgCluster "vianden" is now in the default queue in Luxembourg

We are pleased to announce that the vianden[1] cluster of Luxembourg is now available in the default queue.

Vianden is a cluster of a single node with 8 MI300X AMD GPUs.

The node features:

  • 2x CPU Intel Xeon Platinum 8470 (52 cores/CPU, 2 threads/cores)
  • 8x AMD Instinct MI300X (192 GiB each)
  • 2TiB RAM
  • 1x 3.2TB NVME disk
  • The AMD MI300X GPUs are not supported by Grid'5000 default system (Debian 11). However, one can easily unlock full GPU functionality by deploying the ubuntu2404-rocm environment:

    fluxembourg$ oarsub -t exotic -t deploy -p vianden -I

    fluxembourg$ kadeploy3 -m vianden-1 ubuntu2404-rocm

    More information in the Exotic page.

    This cluster was funded by the University of Luxembourg.

    [1] https://www.grid5000.fr/w/Luxembourg:Hardware#vianden

    -- Grid'5000 Team 11:30, 27 June 2025 (CEST)

    Rss.svgCluster "hydra" is now in the default queue in Lyon

    We are pleased to announce that the hydra[1] cluster of Lyon is now available in the default queue.

    As a reminder, Hydra is a cluster composed of 4 NVIDIA Grace-Hopper servers[2].

    Each node features:

  • 1 Nvidia Grace ARM64 CPU with 72 cores (Neoverse-V2)
  • 1 Nvidia Hopper GPU
  • 512GB LPDDR5 memory
  • 96GB HBM memory
  • 1x1To SSD NVME + 1x1.92To SCSI disk
  • Due to its bleeding-edge hardware, the usual Grid'5000 environments are not supported by default for this cluster.

    (Hydra requires system environments featuring a Linux kernel >= 6.6). The default system on the hydra nodes is based on Debian 11, but **does not provide functional GPUs**. However, users may deploy the ubuntugh2404-arm64-big environment, which is similar to the official Nvidia image provided for this machine and provides GPU support.

    To submit a job on this cluster, the following command may be used:

    oarsub -t exotic -p hydra

    This cluster is funded by INRIA and by Laboratoire de l'Informatique du Parallélisme with ENS Lyon support.

    [1] Hydra is the largest of the modern constellations according to Wikipedia: https://en.wikipedia.org/wiki/Hydra_(constellation)

    [2] https://developer.nvidia.com/blog/nvidia-grace-hopper-superchip-architecture-in-depth/

    -- Grid'5000 Team 16:42, 12 June 2025 (CEST)

    Rss.svgCluster "estats" (Jetson nodes in Toulouse) is now kavlan capable

    The network topology of the estats Jetson nodes can now be configured, just like for other clusters.

    More info in the Network reconfiguration tutorial.

    -- Grid'5000 Team 18:25, 21 May 2025 (CEST)

    Rss.svgCluster "chirop" is now in the default queue of Lille with energy monitoring.

    Dear users,

    We are pleased to announce that the Chirop[1] cluster of Lille is now available in the default queue.

    This cluster consists of 5 HPE DL360 Gen10+ nodes with:

  • 2 CPU Intel Xeon Platinum 8358 (32 cores per CPU)
  • 512 GiB memory
  • 1*1.92TB SSD NVME + 2*3.84TB SSD
  • 2*25 Gbps Ethernet interface
  • Energy monitoring[2] is also available for this cluster[3], provided by newly installed Wattmetres (similar to those already available at Lyon).

    This cluster was funded by CPER CornelIA.

    [1] https://www.grid5000.fr/w/Lille:Hardware#chirop

    [2] https://www.grid5000.fr/w/Energy_consumption_monitoring_tutorial [3] https://www.grid5000.fr/w/Monitoring_Using_Kwollect#Metrics_available_in_Grid.275000

    -- Grid'5000 Team 16:25, 05 May 2025 (CEST)


    Read more news

    Grid'5000 sites

    Current funding

    INRIA

    Logo INRIA.gif

    CNRS

    CNRS-filaire-Quadri.png

    Universities

    IMT Atlantique
    Université Grenoble Alpes, Grenoble INP
    Université Rennes 1, Rennes
    Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
    Université Bordeaux 1, Bordeaux
    Université Lille 1, Lille
    École Normale Supérieure, Lyon

    Regional councils

    Aquitaine
    Auvergne-Rhône-Alpes
    Bretagne
    Champagne-Ardenne
    Provence Alpes Côte d'Azur
    Hauts de France
    Lorraine