Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and AI.
- provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
- highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
- advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
- designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
- a vibrant community of 500+ users supported by a solid technical team
Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.
Grid'5000 is merging with FIT to build the SILECS Infrastructure for Large-scale Experimental Computer Science. Read an Introduction to SILECS (April 2018)
Recently published documents and presentations:
Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).
Current status (at 2022-12-04 14:03)
: 3 current events, None planned (details)
Random pick of publications
Five random publications that benefited from Grid'5000 (at least 2383 overall):
- Thierry Garcia, Pierre Spitéri, Lilia Ziane Khodja, Raphael Couturier. Coupling parallel asynchronous multisplitting methods with Krylov methods to solve pseudo-linear evolution 3D problems. Journal of computational science, 2021, 51, pp.101303. 10.1016/j.jocs.2021.101303. hal-03186558 view on HAL pdf
- Davaadorj Battulga, Daniele Miorandi, Cédric Tedeschi. SpecK: Composition of Stream Processing Applications over Fog Environments. DAIS 2021 - 21st International Conference on Distributed Applications and Interoperable Systems, Jun 2021, Valetta, Malta. pp.38-54, 10.1007/978-3-030-78198-9_3. hal-03259975 view on HAL pdf
- Francesca Ronchini, Romain Serizel, Nicolas Turpault, Samuele Cornell. The impact of non-target events in synthetic soundscapes for sound event detection. DCASE 2021 - Detection and Classification of Acoustic Scenes and Events, Nov 2021, Barcelona/Virtual, Spain. hal-03355184 view on HAL pdf
- Jose Jurandir Alves Esteves, Amina Boubendir, Fabrice Guillemin, Pierre Sens. DRL-based Slice Placement Under Non-Stationary Conditions. CNSM 2021 - 17th International Conference on Network and Service Management, Oct 2021, Izmir, Turkey. hal-03332502 view on HAL pdf
- Zeina Houmani. Data-driven Management Solution for Microservice-based Deep Learning Applications. Other cs.OH. Université de Lyon, 2021. English. NNT : 2021LYSEN092. tel-03593003 view on HAL pdf
New documentation for the FPGA nodes and wattmeter update in Grenoble
Documentation is now available at FPGA for the use of the FPGA nodes (servan cluster in Grenoble).
Furthermore, the Servan nodes' power supply units are now monitored by the Grenoble wattmeter and available in Kwollect.
Finally, due to the unreliability of the measures, Dahu is not monitored by the wattmeter anymore. Past problems with Yeti and Troll should now be fixed.
-- Grid'5000 Team 14:20, Sept 15th 2022 (CEST)
Shorter command usage for Kadeploy tools
We changed the syntax of the Kadeploy commands. A typical command line for deployment was:
kadeploy3 -f $OAR_NODE_FILE -e debian11-x64-base -k
-e flag can now be omitted when deploying a recorded environment, while when deploying an anonymous deployment, the
-a flag must still be set.
- The public key is copied by default,
-k can be omitted.
- Nodes to be deployed are taken from
$OAR_NODE_FILE if not specified on the command line with
As a result, the same deployment can be achieved with the following shorter command line:
Moreover, the architectures are removed from the Grid'5000 reference environments names (eg: "debian11-x64-base" and "debian11-ppc64-base" both become "debian11-base"). Kadeploy infers the correct environment (CPU architecture) for the cluster to deploy. As a result, the previous kadeploy command line can be shortened to:
Old syntaxes and environment names including the architecture are however still supported for backward compatibility.
You can find more information on the Getting Started and Advanced Kadeploy pages.
-- Grid'5000 Team 10:50, Wed, May 19th 2022 (CEST)
Cluster "sirius" from Nvidia is available in the default queue
We have the pleasure to announce that a new cluster named "sirius" is available at Lyon.
This cluster consists in only one Nvidia DGX A100 node with 2 AMD EPYC 7742 CPUs (64 cores per CPU) and 8 Nvidia A100 (40 GiB) GPUs, 1TB of DDR4 RAM, 2x1.92 TB SSD + 4x3.84 TB SSD disks and a coming-soon infiniband network.
Energy monitoring is available for this cluster, provided by the same wattmeter devices as used for the other clusters in Lyon.
This cluster is tagged as "exotic", so the `-t exotic` option must be provided to oarsub to select sirius.
This machine has been funded by LIP laboratory with the support of INRIA Grenoble Rhone-Alpes and ENS Lyon.
-- Grid'5000 Team 11:30, Wed, May 11th 2022 (CEST)
Read more news
As from June 2008, Inria is the main contributor to Grid'5000 funding.
Université Grenoble Alpes, Grenoble INP
Université Rennes 1, Rennes
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
Université Bordeaux 1, Bordeaux
Université Lille 1, Lille
École Normale Supérieure, Lyon
Provence Alpes Côte d'Azur
Hauts de France