Grid5000:Home: Difference between revisions

From Grid5000
Jump to navigation Jump to search
No edit summary
No edit summary
 
(163 intermediate revisions by 13 users not shown)
Line 1: Line 1:
__NOTOC__ __NOEDITSECTION__
__NOTOC__ __NOEDITSECTION__
{|width="95%"
{|width="95%"
|-
|- valign="top"
| width="20%" |
|bgcolor="#888888" style="border:1px solid #cccccc;padding:2em;padding-top:1em;"|
[[Image:Logo_Aladdin.png|250px]]
[[File:Slices-ri-white-color.png|260px|left]]
|
<b>Grid'5000 is a precursor infrastructure of [http://www.slices-ri.eu SLICES-RI], Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.</b>
= ALADDIN-G5K : ensuring the development of '''Grid'5000''' =
<br/>
= for the 2008-2011 period =
Content on this website is partly outdated. Technical information remains relevant.
|}
 
{|width="95%"
|- valign="top"
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
[[Image:g5k-backbone.png|thumbnail|260px|right|Grid'5000]]
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.'''
 
Key features:
* provides '''access to a large amount of resources''': 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
* '''designed to support Open Science and reproducible research''', with full traceability of infrastructure and software changes on the testbed
* '''a vibrant community''' of 500+ users supported by a solid technical team
 
<br>
Read more about our [[Team|teams]], our [[Publications|publications]], and the [[Grid5000:UsagePolicy|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].


''An infrastructure distributed in 9 sites around France, for research in large-scale parallel and distributed systems''
<br>
Published documents and presentations:
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]


Engineers ensuring the development and day to day support of the infrastructure are mostly provided by Inria, under the ''ADT ALADDIN-G5K'' initiative.
Older documents:
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]


<br>
Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL [[Hemera|HEMERA]] (2010-2014).
|}
|}


{|width="95%"
<br>
|- valign="top"
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
<br>
 
== Random pick of publications ==
{{#publications:}}
 
==Latest news==
==Latest news==
[[Image:Logo_Aladdin.png|left|84px|ALADDIN-G5K]]
<rss max=4 item-max-length="2000">https://www.grid5000.fr/rss/G5KNews.php</rss>
=== Registration for Grid'5000 spring school now open ===
The Grid'5000 spring school is sheduled between April 7, 2009 and April 10, 2009 in Nancy. A [[Grid5000:School2009|dedicated page]] has been setup, where you can find the registration instructions, guiding you to this link: http://registration.net-resa.com/cgi-bin/WebObjects/gnetresa.woa/portal/org/48/evt/364/ .
=== Latest updated experiment descriptions ===
{{#experiments:3}}
----
=== Latest updated publications ===
{{#publications:3}}
----
[[Image:People_Workshop_CEA_1.jpeg|right|84px|Photo of people present]]
=== First INRIA-CEA workshop held on deployment and scheduling in large infrastructures ===
This workshop, held noon to noon January 22nd-23rd discussed common problems faced when dealing with deployment and scheduling in large infrastructures. Most participants are visible in the photo on the right of this page.
[[Image:Logo_Aladdin.png|left|84px|ALADDIN-G5K]]
=== Grid'5000 spring school to be held in Nancy, in April 2009 ===
A Grid'5000 spring school is in preparation for 4 days between April 7, 2009 and April 10, 2009 in Nancy. A [[Grid5000:School2009|dedicated page]] has been setup, which includes a [[Grid5000:School2009#Call_for_presentation_and_practical_sessions|call for presentations and practical sessions]].
----
----
[[Grid5000:News|read more news]]
[[News|Read more news]]
|}


<br>
=== Grid'5000 sites===
==Grid'5000 at a glance==
{|width="100%" cellspacing="3"  
[[Image:site_map.png|thumbnail|128px|right|Grid'5000 sites]]
* '''Grid'5000''' project aims at building a '''highly reconfigurable, controlable and monitorable experimental Grid platform''' gathering '''9 sites''' geographically distributed in France featuring a total of 5000 processors:
===Sites:===
{|width="75%" cellspacing="3"  
|- valign="top"
|- valign="top"
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Bordeaux:Home|Bordeaux]]
* [[Grenoble:Home|Grenoble]]
* [[Grenoble:Home|Grenoble]]
* [[Lille:Home|Lille]]
* [[Lille:Home|Lille]]
* [[Luxembourg:Home|Luxembourg]]
* [[Louvain:Home|Louvain]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Lyon:Home|Lyon]]
* [[Lyon:Home|Lyon]]
* [[Nancy:Home|Nancy]]
* [[Nancy:Home|Nancy]]
* [[Orsay:Home|Orsay]]
* [[Nantes:Home|Nantes]]
* [[Rennes:Home|Rennes]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* [[Rennes:Home|Rennes]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Sophia:Home|Sophia-Antipolis]]
* [[Strasbourg:Home|Strasbourg]]
* [[Toulouse:Home|Toulouse]]
* [[Toulouse:Home|Toulouse]]
|-
|-
|}
|}


* The main purpose of this platform is to serve as an experimental testbed for research in Grid Computing.
== Current funding ==
* This project is one initiative of the [http://www.recherche.gouv.fr/recherche/fns/grid.htm French ACI Grid] incentive (see below: Funding Institutions) which provides a large part of Grid'5000 funding on behalf of the French Ministry of Research & Education.
 
 
 
[[Image:Software layers.png|thumbnail|271px|left|Grid'5000 will allow Grid experiments France wide in all these software layers]]
* '''Grid'5000''' is a research effort developping a '''large scale nation wide infrastructure for Grid research'''.
 
* '''17 [[Grid5000:Laboratories|laboratories]]''' are involved, nation wide, in the objective of providing the community of Grid researchers a testbed allowing experiments in all the software layers between the network protocols up to the applications.
 
 
 
The current plans are to assemble a physical platform featuring 9 local platform (at least one cluster per site), each with 100 to a thousand PCs, connected by the [http://www.renater.fr RENATER] Education and Research Network.
 
All clusters will be connected to Renater with a 10Gb/s link (or at least 1 Gb/s, when 10Gb/s is not available yet).
 
 
 
This high collaborative research effort is funded by the French ministry of Education and Research, INRIA, CNRS, the Universities of all sites and some regional councils.
 
 
 
==Rationale==
'''The foundations of Grid'5000''' have emerged from a thorough analysis and numerous discussions about methodologies used for scientific research in the Grid domain. A report presents the [http://www-sop.inria.fr/aci/grid/public/Library/rapport-grid5000-V3.pdf rationale for Grid'5000].
 
In addition to theory, simulators and emulators, there is a strong need for '''large scale testbeds''' where real life experimental conditions hold. '''The size of Grid'5000''', in terms of number of sites and number of processors per site, was established according to the scale of the experiments and the number of researchers involved in the project.
 
 
 
==Funding Institutions==
{|width="100%" cellspacing="3"
{|width="100%" cellspacing="3"
|-
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Ministère de l'Education, de la Jeunesse et de la Recherche===
[[Image:Logo-ministere.png]]
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===ACI Grid===
[[Image:LogoACIGRID.jpg|350px]]
|-
|-
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===INRIA===
===INRIA===
[[Image:Logo-inria.png]]
[[Image:Logo_INRIA.gif|300px]]
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===CNRS===
===CNRS===
[[Image:Logo-cnrs.png]]
[[Image:CNRS-filaire-Quadri.png|125px]]
|-
|-
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Universities===
===Universities===
University of Paris Sud, Orsay<br/>
IMT Atlantique<br/>
University Joseph Fourier, Grenoble<br/>
Université Grenoble Alpes, Grenoble INP<br/>
University of Nice-Sophia Antipolis, Sophia Antipolis<br/>
Université Rennes 1, Rennes<br/>
University of Rennes 1, Rennes<br/>
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
University Bordeaux 1, Bordeaux<br/>
Université Bordeaux 1, Bordeaux<br/>
University Lille 1 / GENOPOLE, Lille<br/>
Université Lille 1, Lille<br/>
Ecole Normale Supérieure / MYRICOM, Lyon<br/>
École Normale Supérieure, Lyon<br/>
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
===Regional councils===
===Regional councils===
Aquitaine<br/>
Auvergne-Rhône-Alpes<br/>
Bretagne<br/>
Bretagne<br/>
Champagne-Ardenne<br/>
Provence Alpes Côte d'Azur<br/>
Provence Alpes Côte d'Azur<br/>
Aquitaine<br/>
Hauts de France<br/>
Ile de France<br/>
Lorraine<br/>
Lorraine<br/>
===General Councils===
Alpes Maritimes
|}
|}

Latest revision as of 09:50, 10 June 2025

Slices-ri-white-color.png

Grid'5000 is a precursor infrastructure of SLICES-RI, Scientific Large Scale Infrastructure for Computing/Communication Experimental Studies.
Content on this website is partly outdated. Technical information remains relevant.

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing, including Cloud, HPC, Big Data and AI.

Key features:

  • provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.


Published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2025-06-22 17:58): 2 current events, 2 planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2774 overall):

  • Charles Bouillaguet, Ambroise Fleury, Pierre-Alain Fouque, Paul Kirchner. We are on the same side. Alternative sieving strategies for the number field sieve. ASIACRYPT 2023 - 29th International Conference on the Theory and Application of Cryptology and Information Security, Dec 2023, Guangzhou, China. pp.138-166, 10.1007/978-981-99-8730-6_5. hal-04112671 view on HAL pdf
  • Céline Acary-Robert, Emmanuel Agullo, Ludovic Courtès, Marek Felšöci, Konrad Hinsen, et al.. Guix-HPC Activity Report 2022–2023. Inria Bordeaux - Sud Ouest. 2024, pp.1-32. hal-04500140 view on HAL pdf
  • Romain Xu-Darme, Julien Girard-Satabin, Darryl Hond, Gabriele Incorvaia, Zakaria Chihani. Contextualised Out-of-Distribution Detection using Pattern Identification. Computer Safety, Reliability, and Security. SAFECOMP 2023 Workshops, Sep 2023, Toulouse, France. cea-04254022 view on HAL pdf
  • Josefine Umlauft, Christopher W. Johnson, Philippe Roux, Daniel Taylor Trugman, Albanne Lecointre, et al.. Mapping Glacier Basal Sliding Applying Machine Learning. Journal of Geophysical Research: Earth Surface, 2023, 128 (11), 10.1029/2023JF007280. insu-04604354 view on HAL pdf
  • Maxime Agusti, Eddy Caron, Benjamin Fichel, Laurent Lefèvre, Olivier Nicol, et al.. PowerHeat: A non-intrusive approach for estimating the power consumption of bare metal water-cooled servers. 2024 IEEE International Conferences on Internet of Things (iThings) and IEEE Green Computing & Communications (GreenCom) and IEEE Cyber, Physical & Social Computing (CPSCom) and IEEE Smart Data (SmartData) and IEEE Congress on Cybermatics, Aug 2024, Copenhagen, Denmark. pp.1-7. hal-04662683 view on HAL pdf


Latest news

Rss.svgCluster "hydra" is now in the default queue in Lyon

We are pleased to announce that the hydra[1] cluster of Lyon is now available in the default queue.

As a reminder, Hydra is a cluster composed of 4 NVIDIA Grace-Hopper servers[2].

Each node features:

  • 1 Nvidia Grace ARM64 CPU with 72 cores (Neoverse-V2)
  • 1 Nvidia Hopper GPU
  • 512GB LPDDR5 memory
  • 96GB HBM memory
  • 1x1To SSD NVME + 1x1.92To SCSI disk
  • Due to its bleeding-edge hardware, the usual Grid'5000 environments are not supported by default for this cluster.

    (Hydra requires system environments featuring a Linux kernel >= 6.6). The default system on the hydra nodes is based on Debian 11, but **does not provide functional GPUs**. However, users may deploy the ubuntugh2404-arm64-big environment, which is similar to the official Nvidia image provided for this machine and provides GPU support.

    To submit a job on this cluster, the following command may be used:

    oarsub -t exotic -p hydra

    This cluster is funded by INRIA and by Laboratoire de l'Informatique du Parallélisme with ENS Lyon support.

    [1] Hydra is the largest of the modern constellations according to Wikipedia: https://en.wikipedia.org/wiki/Hydra_(constellation)

    [2] https://developer.nvidia.com/blog/nvidia-grace-hopper-superchip-architecture-in-depth/

    -- Grid'5000 Team 16:42, 12 June 2025 (CEST)

    Rss.svgCluster "estats" (Jetson nodes in Toulouse) is now kavlan capable

    The network topology of the estats Jetson nodes can now be configured, just like for other clusters.

    More info in the Network reconfiguration tutorial.

    -- Grid'5000 Team 18:25, 21 May 2025 (CEST)

    Rss.svgCluster "chirop" is now in the default queue of Lille with energy monitoring.

    Dear users,

    We are pleased to announce that the Chirop[1] cluster of Lille is now available in the default queue.

    This cluster consists of 5 HPE DL360 Gen10+ nodes with:

  • 2 CPU Intel Xeon Platinum 8358 (32 cores per CPU)
  • 512 GiB memory
  • 1*1.92TB SSD NVME + 2*3.84TB SSD
  • 2*25 Gbps Ethernet interface
  • Energy monitoring[2] is also available for this cluster[3], provided by newly installed Wattmetres (similar to those already available at Lyon).

    This cluster was funded by CPER CornelIA.

    [1] https://www.grid5000.fr/w/Lille:Hardware#chirop

    [2] https://www.grid5000.fr/w/Energy_consumption_monitoring_tutorial [3] https://www.grid5000.fr/w/Monitoring_Using_Kwollect#Metrics_available_in_Grid.275000

    -- Grid'5000 Team 16:25, 05 May 2025 (CEST)

    Rss.svgChange of default queue based on platform

    Until now, Abaca (production) users had to specify `-q production` when reserving Abaca resources with OAR.

    This is no longer necessary as your default queue is now automatically selected based on the platform your default group is associated to, as shown at https://api.grid5000.fr/explorer/selector/ and in the message displayed when connecting to a frontend.

    For SLICES-FR users, there is no change since the correct queue was already selected by default.

    Additionally, the "production" queue has been renamed to "abaca", although "production" will continue to work for the foreseeable future.

    Please note one case where this change may affect your workflow:

    When an Abaca user reserves a resource from SLICES-FR (a non-production resource), they must explicitly specify they want to use the SLICES-FR queue, which is called "default", by adding `-q default` the OAR command.

    -- Abaca Grid'5000 Team 10:10, 31 March 2025 (CEST)


    Read more news

    Grid'5000 sites

    Current funding

    INRIA

    Logo INRIA.gif

    CNRS

    CNRS-filaire-Quadri.png

    Universities

    IMT Atlantique
    Université Grenoble Alpes, Grenoble INP
    Université Rennes 1, Rennes
    Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
    Université Bordeaux 1, Bordeaux
    Université Lille 1, Lille
    École Normale Supérieure, Lyon

    Regional councils

    Aquitaine
    Auvergne-Rhône-Alpes
    Bretagne
    Champagne-Ardenne
    Provence Alpes Côte d'Azur
    Hauts de France
    Lorraine