Grid5000:Home: Difference between revisions

From Grid5000
Jump to navigation Jump to search
No edit summary
No edit summary
 
(25 intermediate revisions by 7 users not shown)
Line 1: Line 1:
__NOTOC__ __NOEDITSECTION__
__NOTOC__ __NOEDITSECTION__
{|width="95%"
|- valign="top"
|bgcolor="#888888" style="border:1px solid #cccccc;padding:2em;padding-top:1em;"|
[[File:Slices-ri-white-color.png|260px|left]]
<b>Grid'5000 is a precursor infrastructure of [http://www.slices-ri.eu SLICES-RI], Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.</b>
<br/>
Content on this website is partly outdated. Technical information remains relevant.
|}
{|width="95%"
{|width="95%"
|- valign="top"
|- valign="top"
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
[[Image:renater5-g5k.jpg|thumbnail|250px|right|Grid'5000]]
[[Image:g5k-backbone.png|thumbnail|260px|right|Grid'5000]]
'''Grid'5000 is a large-scale and versatile testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data.'''
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.'''


Key features:
Key features:
* provides '''access to a large amount of resources''': 1000 nodes, 8000 cores, grouped in homogeneous clusters, and featuring various technologies: 10G Ethernet, Infiniband, GPUs, Xeon PHI
* provides '''access to a large amount of resources''': 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
Line 17: Line 26:


<br>
<br>
Recently published documents:
Published documents and presentations:
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]
 
Older documents:
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board meeting (2014)]
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]


<br>
<br>
Line 28: Line 41:
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
<br>
<br>
== Random pick of publications ==
{{#publications:}}


==Latest news==
==Latest news==
<rss max=4 item-max-length="2000">https://www.grid5000.fr/mediawiki/index.php?title=News&action=feed&feed=atom</rss>
<rss max=4 item-max-length="2000">https://www.grid5000.fr/rss/G5KNews.php</rss>
----
----
[[News|Read more news]]
[[News|Read more news]]
Line 41: Line 57:
* [[Lille:Home|Lille]]
* [[Lille:Home|Lille]]
* [[Luxembourg:Home|Luxembourg]]
* [[Luxembourg:Home|Luxembourg]]
* [[Louvain:Home|Louvain]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Lyon:Home|Lyon]]
* [[Lyon:Home|Lyon]]
* [[Nancy:Home|Nancy]]
* [[Nancy:Home|Nancy]]
* [[Nantes:Home|Nantes]]
* [[Nantes:Home|Nantes]]
* [[Rennes:Home|Rennes]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Rennes:Home|Rennes]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Strasbourg:Home|Strasbourg]]
* [[Toulouse:Home|Toulouse]]
* [[Toulouse:Home|Toulouse]]
|-
|-
Line 53: Line 71:


== Current funding ==
== Current funding ==
As from June 2008, Inria is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].
{|width="100%" cellspacing="3"
{|width="100%" cellspacing="3"
|-
|-
Line 65: Line 82:
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Universities===
===Universities===
IMT Atlantique<br/>
Université Grenoble Alpes, Grenoble INP<br/>
Université Grenoble Alpes, Grenoble INP<br/>
Université Rennes 1, Rennes<br/>
Université Rennes 1, Rennes<br/>
Line 74: Line 92:
===Regional councils===
===Regional councils===
Aquitaine<br/>
Aquitaine<br/>
Auvergne-Rhône-Alpes<br/>
Bretagne<br/>
Bretagne<br/>
Champagne-Ardenne<br/>
Champagne-Ardenne<br/>
Provence Alpes Côte d'Azur<br/>
Provence Alpes Côte d'Azur<br/>
Nord Pas de Calais<br/>
Hauts de France<br/>
Lorraine<br/>
Lorraine<br/>
|}
|}

Latest revision as of 09:50, 10 June 2025

Slices-ri-white-color.png

Grid'5000 is a precursor infrastructure of SLICES-RI, Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.
Content on this website is partly outdated. Technical information remains relevant.

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.

Key features:

  • provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.


Published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2025-07-01 12:39): 1 current events, 7 planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2777 overall):

  • Adrien Schoen, Gregory Blanc, Pierre-François Gimenez, Yufei Han, Frédéric Majorczyk, et al.. A tale of two methods: unveiling the limitations of GAN and the rise of bayesian networks for synthetic network traffic generation. 2024 IEEE European Symposium on Security and Privacy Workshops (EuroS&PW), Jul 2024, Vienna, Austria. pp.273-286, 10.1109/EuroSPW61312.2024.00036. hal-04871298 view on HAL pdf
  • Josefine Umlauft, Christopher W. Johnson, Philippe Roux, Daniel Taylor Trugman, Albanne Lecointre, et al.. Mapping Glacier Basal Sliding Applying Machine Learning. Journal of Geophysical Research: Earth Surface, 2023, 128 (11), 10.1029/2023JF007280. insu-04604354 view on HAL pdf
  • Barbara Gendron, Gaël Guibon. SEC: Context-Aware Metric Learning for Efficient Emotion Recognition in Conversation. Proceedings of the 14th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis (WASSA at ACL 2024), Aug 2024, Bangkok, Thailand. hal-04702997 view on HAL pdf
  • Luca Di Stefano, Frédéric Lang. Compositional verification of priority systems using sharp bisimulation. Formal Methods in System Design, 2024, 62, pp.1-40. 10.1007/s10703-023-00422-1. hal-04103681 view on HAL pdf
  • Céline Acary-Robert, Emmanuel Agullo, Ludovic Courtès, Marek Felšöci, Konrad Hinsen, et al.. Guix-HPC Activity Report 2022–2023. Inria Bordeaux - Sud Ouest. 2024, pp.1-32. hal-04500140 view on HAL pdf


Latest news

Rss.svgCluster "vianden" is now in the default queue in Luxembourg

We are pleased to announce that the vianden[1] cluster of Luxembourg is now available in the default queue.

Vianden is a cluster of a single node with 8 MI300X AMD GPUs.

The node features:

  • 2x CPU Intel Xeon Platinum 8470 (52 cores/CPU, 2 threads/cores)
  • 8x AMD Instinct MI300X (192 GiB each)
  • 2TiB RAM
  • 1x 3.2TB NVME disk
  • The AMD MI300X GPUs are not supported by Grid'5000 default system (Debian 11). However, one can easily unlock full GPU functionality by deploying the ubuntu2404-rocm environment:

    fluxembourg$ oarsub -t exotic -t deploy -p vianden -I

    fluxembourg$ kadeploy3 -m vianden-1 ubuntu2404-rocm

    More information in the Exotic page.

    This cluster was funded by the University of Luxembourg.

    [1] https://www.grid5000.fr/w/Luxembourg:Hardware#vianden

    -- Grid'5000 Team 11:30, 27 June 2025 (CEST)

    Rss.svgCluster "hydra" is now in the default queue in Lyon

    We are pleased to announce that the hydra[1] cluster of Lyon is now available in the default queue.

    As a reminder, Hydra is a cluster composed of 4 NVIDIA Grace-Hopper servers[2].

    Each node features:

  • 1 Nvidia Grace ARM64 CPU with 72 cores (Neoverse-V2)
  • 1 Nvidia Hopper GPU
  • 512GB LPDDR5 memory
  • 96GB HBM memory
  • 1x1To SSD NVME + 1x1.92To SCSI disk
  • Due to its bleeding-edge hardware, the usual Grid'5000 environments are not supported by default for this cluster.

    (Hydra requires system environments featuring a Linux kernel >= 6.6). The default system on the hydra nodes is based on Debian 11, but **does not provide functional GPUs**. However, users may deploy the ubuntugh2404-arm64-big environment, which is similar to the official Nvidia image provided for this machine and provides GPU support.

    To submit a job on this cluster, the following command may be used:

    oarsub -t exotic -p hydra

    This cluster is funded by INRIA and by Laboratoire de l'Informatique du Parallélisme with ENS Lyon support.

    [1] Hydra is the largest of the modern constellations according to Wikipedia: https://en.wikipedia.org/wiki/Hydra_(constellation)

    [2] https://developer.nvidia.com/blog/nvidia-grace-hopper-superchip-architecture-in-depth/

    -- Grid'5000 Team 16:42, 12 June 2025 (CEST)

    Rss.svgCluster "estats" (Jetson nodes in Toulouse) is now kavlan capable

    The network topology of the estats Jetson nodes can now be configured, just like for other clusters.

    More info in the Network reconfiguration tutorial.

    -- Grid'5000 Team 18:25, 21 May 2025 (CEST)

    Rss.svgCluster "chirop" is now in the default queue of Lille with energy monitoring.

    Dear users,

    We are pleased to announce that the Chirop[1] cluster of Lille is now available in the default queue.

    This cluster consists of 5 HPE DL360 Gen10+ nodes with:

  • 2 CPU Intel Xeon Platinum 8358 (32 cores per CPU)
  • 512 GiB memory
  • 1*1.92TB SSD NVME + 2*3.84TB SSD
  • 2*25 Gbps Ethernet interface
  • Energy monitoring[2] is also available for this cluster[3], provided by newly installed Wattmetres (similar to those already available at Lyon).

    This cluster was funded by CPER CornelIA.

    [1] https://www.grid5000.fr/w/Lille:Hardware#chirop

    [2] https://www.grid5000.fr/w/Energy_consumption_monitoring_tutorial [3] https://www.grid5000.fr/w/Monitoring_Using_Kwollect#Metrics_available_in_Grid.275000

    -- Grid'5000 Team 16:25, 05 May 2025 (CEST)


    Read more news

    Grid'5000 sites

    Current funding

    INRIA

    Logo INRIA.gif

    CNRS

    CNRS-filaire-Quadri.png

    Universities

    IMT Atlantique
    Université Grenoble Alpes, Grenoble INP
    Université Rennes 1, Rennes
    Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
    Université Bordeaux 1, Bordeaux
    Université Lille 1, Lille
    École Normale Supérieure, Lyon

    Regional councils

    Aquitaine
    Auvergne-Rhône-Alpes
    Bretagne
    Champagne-Ardenne
    Provence Alpes Côte d'Azur
    Hauts de France
    Lorraine