Difference between revisions of "Grid5000:Home"

From Grid5000
Jump to: navigation, search
(Fix news link)
 
(297 intermediate revisions by 16 users not shown)
Line 1: Line 1:
 
__NOTOC__ __NOEDITSECTION__
 
__NOTOC__ __NOEDITSECTION__
 
 
{|width="95%"
 
{|width="95%"
 
|- valign="top"
 
|- valign="top"
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
[[Image:Logo_sc05.gif|64px|left|SC2005]]
+
[[Image:renater5-g5k.jpg|thumbnail|250px|right|Grid'5000]]
==Breaking news==
+
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and AI.'''
* '''Grid'5000 @ SC2005''': Grid'5000 platform demonstations are presented at SuperComputing (see [[Grid5000:News|here]]).
+
 
 +
Key features:
 +
* provides '''access to a large amount of resources''': 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
 +
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
 +
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
 +
* '''designed to support Open Science and reproducible research''', with full traceability of infrastructure and software changes on the testbed
 +
* '''a vibrant community''' of 500+ users supported by a solid technical team
 +
 
 +
<br>
 +
Read more about our [[Team|teams]], our [[Publications|publications]], and the [[Grid5000:UsagePolicy|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].
 +
 
 +
<b>Grid'5000 is merging with [https://fit-equipex.fr FIT] to build the [http://www.silecs.net/ SILECS Infrastructure for Large-scale Experimental Computer Science]. Read [http://www.silecs.net/wp-content/uploads/2018/04/Desprez-SILECS.pdf an Introduction to SILECS] (April 2018)</b>
 +
 
 +
<br>
 +
Recently published documents and presentations:
 +
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
 +
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]
 +
 
 +
Older documents:
 +
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
 +
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]
 +
 
 +
<br>
 +
Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL [[Hemera|HEMERA]] (2010-2014).
 
|}
 
|}
  
==Grid'5000 at a glance==
+
<br>
[[Image:CarteSites.JPG|thumbnail|128px|right|Grid'5000 sites]]
+
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
* '''Grid'5000''' project aims at building an '''experimental Grid platform''' gathering '''9 sites''' geographically distributed in France:
+
<br>
===Sites:===
+
 
{|width="75%" cellspacing="3"  
+
== Random pick of publications ==
 +
{{#publications:}}
 +
 
 +
==Latest news==
 +
<rss max=4 item-max-length="2000">https://www.grid5000.fr/w?title=News&action=feed&feed=atom</rss>
 +
----
 +
[[News|Read more news]]
 +
 
 +
=== Grid'5000 sites===
 +
{|width="100%" cellspacing="3"  
 
|- valign="top"
 
|- valign="top"
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Bordeaux
+
* [[Grenoble:Home|Grenoble]]
* Grenoble
+
* [[Lille:Home|Lille]]
* Lille
+
* [[Luxembourg:Home|Luxembourg]]
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Lyon
+
* [[Lyon:Home|Lyon]]
* Nancy
+
* [[Nancy:Home|Nancy]]
* Orsay
+
* [[Nantes:Home|Nantes]]
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Rennes
+
* [[Rennes:Home|Rennes]]
* Sophia-Antipolis
+
* [[Sophia:Home|Sophia-Antipolis]]
* Toulouse
+
* [[Toulouse:Home|Toulouse]]
 
|-
 
|-
 
|}
 
|}
  
* The main purpose of this platform is to serve as an experimental testbed for research in Grid Computing.
+
== Current funding ==
* This project is one initiative of the '''French ACI Grid''' Incentive [[Image:LogoACIGRID.jpg]]
+
As from June 2008, Inria is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].  
 
 
==Introduction==
 
* '''Grid'5000''' is a research effort developping a '''large scale nation wide infrastructure for Grid research'''.
 
 
 
* '''17 [[Grid5000:Laboratories|laboratories]]''' are involved, nation wide, in the objective of providing the community of Grid researchers a testbed allowing experiments in all the software layers between the network protocols up to the applications:
 
{|
 
|
 
{| border=1
 
|- bgcolor="#EEEEEE"
 
| Applications
 
|- bgcolor="#DDDDDD"
 
| Algorithms
 
|- bgcolor="#CCCCCC"
 
| Runtime
 
|- bgcolor="#BBBBBB"
 
| Middleware
 
|- bgcolor="#AAAAAA"
 
| Operating Systems
 
|- bgcolor="#999999"
 
| Network protocols
 
|}
 
| Grid'5000 will allow Grid experiments France wide in all these software layers
 
|}
 
The current plans are to assemble a physical platform featuring 9 clusters, each with 100 to a thousand PCs, connected by the [http://www.renater.fr RENATER] Education and Research Network. All clusters will be connected to Renater at 1 Gb/s (10 Gb/s is expected in near future).
 
 
 
This high collaborative research effort is funded by the French ministry of Education and Research, INRIA, CNRS, the Universities of all sites and some regional councils.
 
 
 
==Funding Institutions==
 
 
{|width="100%" cellspacing="3"
 
{|width="100%" cellspacing="3"
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Ministère de l'Education, de la Jeunesse et de la Recherche===
 
[[Image:Logo-Ministere-Recherche.jpg]]
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===ACI Grid===
 
[[Image:LogoACIGRID.jpg]]
 
 
|-
 
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===INRIA===
 
===INRIA===
[[Image:Logo-inria.gif]]
+
[[Image:Logo_INRIA.gif|300px]]
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===CNRS===
 
===CNRS===
[[Image:Logo-cnrs.gif]]
+
[[Image:CNRS-filaire-Quadri.png|125px]]
 
|-
 
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Universities===
 
===Universities===
University of Paris Sud, Orsay<br/>
+
Université Grenoble Alpes, Grenoble INP<br/>
University Joseph Fourier, Grenoble<br/>
+
Université Rennes 1, Rennes<br/>
University of Nice-Sophia Antipolis, Sophia Antipolis<br/>
+
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
University of Rennes 1, Rennes<br/>
+
Université Bordeaux 1, Bordeaux<br/>
University Paul Sabatier / FERIA, Toulouse<br/>
+
Université Lille 1, Lille<br/>
University Bordeaux 1, Bordeaux<br/>
+
École Normale Supérieure, Lyon<br/>
University Lille 1 / GENOPOLE, Lille<br/>
 
Ecole Normale Supérieure / MYRICOM, Lyon<br/>
 
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Regional councils===
 
===Regional councils===
 +
Aquitaine<br/>
 +
Auvergne-Rhône-Alpes<br/>
 
Bretagne<br/>
 
Bretagne<br/>
 +
Champagne-Ardenne<br/>
 
Provence Alpes Côte d'Azur<br/>
 
Provence Alpes Côte d'Azur<br/>
Aquitaine<br/>
+
Hauts de France<br/>
Ile de France<br/>
+
Lorraine<br/>
===General Councils===
 
Alpes Maritimes
 
 
|}
 
|}

Latest revision as of 14:30, 12 February 2021

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and AI.

Key features:

  • provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.

Grid'5000 is merging with FIT to build the SILECS Infrastructure for Large-scale Experimental Computer Science. Read an Introduction to SILECS (April 2018)


Recently published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2021-07-27 20:58): 1 current events, None planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2181 overall):

  • Sunit Sivasankaran. Localization guided speech separation. Machine Learning cs.LG. Université de Lorraine, 2020. English. NNT : 2020LORR0078. tel-02961882 view on HAL pdf
  • Oleksii Avilov, Sébastien Rimbert, Anton Popov, Laurent Bougrain. Optimizing Motor Intention Detection with Deep Learning: Towards Management of Intraoperative Awareness. IEEE Transactions on Biomedical Engineering, Institute of Electrical and Electronics Engineers, 2021, pp.10. 10.1109/TBME.2021.3064794. hal-03166495 view on HAL pdf
  • Sara Dahmani, Vincent Colotte, Slim Ouni. Étude comparative des paramètres d'entrée pour la synthèse expressive audiovisuelle de la parole par DNNs. 6e conférence conjointe Journées d'Études sur la Parole (JEP, 33e édition), Traitement Automatique des Langues Naturelles (TALN, 27e édition), Rencontre des Étudiants Chercheurs en Informatique pour le Traitement Automatique des Langues (RÉCITAL, 22e édition). Volume 1 : Journées d'Études sur la Parole, Jun 2020, Nancy, France. pp.127-135. hal-02798526v3 view on HAL pdf
  • Mathias Fleury, Hans-Jörg Schurr. Reconstructing veriT Proofs in Isabelle/HOL. PxTP 2019 - Sixth Workshop on Proof eXchange for Theorem Proving, Aug 2019, Natal, Brazil. pp.36-50, 10.4204/EPTCS.301.6. hal-02276530 view on HAL pdf
  • Manuel Alejandro Diaz-Zapata. Contribution to Panoptic Segmentation. Technical Report RT-0506, Inria; Universidad Autónoma de Occidente. 2019. hal-02300774v2 view on HAL pdf


Latest news

Rss.svgGrid'5000 metadata bundler is available in alpha version

Dear users,

When running experiments on Grid'5000, users generate metadata across multiple services.

The extraction and the storage of this metadata is useful for matters of scientific data management and reproducibility. We are currently developing a tool designed to collect this metadata concerning an experiment from the different Grid'5000 services and bundle it in a single compressed archive. Making it easier to find, study, and store information about your experiments.

The g5k-metadata-bundler is now available in alpha version on all Grid'5000 node frontends.

At this point in time, the bundles generated by g5k-metadata-bundler contains:

  • the OAR information of a single job,
  • copies from the specifications of the nodes involved,
  • all monitoring information collected by kwollect during the job.

As the software is still in alpha version, the contents and structure of the bundle is subject to changes and a list of planned features is available on the dedicated wiki page.

We are also interested in user feedback on the support mailing list.

-- Grid'5000 Team 14:30, July 22nd 2021 (CEST)

Rss.svgNew AMD cluster named "neowise" available in Lyon for testing

Dear users,

A new cluster, named "neowise" is available in Lyon for testing.

This machine is a donation from AMD to Genci and Inria to support the French research community (in particular for works in relation with COVID19).

The cluster has 10 nodes, each including an AMD EPYC 7642 48-Core processor, 512GB of RAM, 8 Radeon MI50 GPUs, and an HDR Infiniband network (200Gbs). Its full characteristics are described at: https://www.grid5000.fr/w/Lyon:Hardware#neowise

The cluster is still in testing phase and few issues are known:

- Few nodes currently have problems and will be unavailable until fixed - The software stack to use AMD GPUs is incomplete. The HIP compiler is included in default environment but many libraries or software (such as Deep Learning frameworks) are lacking. They will be added soon (mostly as "environment modules") - Some Grid'5000 "advanced" features are missing. An overview of what is working correctly is available at: https://intranet.grid5000.fr/jenkins-status?config=neowise

The neowise cluster is tagged as "exotic" and is currently available in "testing" queue. To submit a job, don't forget to add the appropriate options to oarsub. For instance:

$ oarsub -q testing -t exotic -p "cluster = 'neowise'" -I

We would like to thank AMD for this donation and Genci for their successful collaboration in making this machine available in Grid'5000.

-- Grid'5000 Team 14:30, June 24th 2021 (CEST)

Rss.svgDebian 11 "Bullseye" preview environments are now available for deployments

Debian 11 stable (Bullseye) will be released in a few weeks. We are pleased to offer a "preview" of kadeploy environments for Debian 11 (currently still Debian testing), that you can deploy already. See the debian11 environments in kaenv3.

New features and changes in Debian 11 are described in: https://www.debian.org/releases/bullseye/amd64/release-notes/ch-whats-new.en.html .

In particular, it includes many software updates:

  • Cuda 11.2 / Nvidia drivers 460.73.01
  • OpenJDK 17
  • Python 3.9.2
    • python3-numpy 1.19.5
    • python3-scipy 1.6.0
    • python3-pandas 1.1.5

(Note that Python 2 is not included as deprecated since Jan. 2020 and /usr/bin/python is symlinked to /usr/bin/python3)

  • Perl 5.32.1
  • GCC 9.3 and 10.2
  • G++ 10.2
  • Libboost 1.74.0
  • Ruby 2.7.3
  • CMake 3.18.4
  • GFortran 10.2.1
  • Liblapack 3.9.0
  • libatlas 3.10.3
  • RDMA 33.1
  • OpenMPI 4.1.0

Known regressions and problems are :

  • The std environment is not ready yet, and will not be the default Grid'5000 environment until official Debian 11 Bullseye release.
  • Cuda/Nvidia drivers do not support - out-of-the-box or at all - some quite old GPUs.
  • BeeGFS is not operational at the moment.

Let us know if you want us to support some tools, softwares,… that are not available on big images.

As a reminder, you can use the following commands to deploy an environment on nodes (https://www.grid5000.fr/w/Getting_Started#Deploying_nodes_with_Kadeploy):

 $ oarsub -t deploy -I
 $ kadeploy3 -e debian11-x64-big...

Rss.svgKadeploy: use of UUID partition identifiers and faster deployments

Up to now, kadeploy was identifying disk partitions with their block device names (e.g. /dev/sda3) when deploying a system. This no longer works reliably because of disk inversion issues with recent kernels. As a result, we have changed kadeploy to use filesystem UUIDs instead.

This change affects the root partition passed to the kernel command line as well as the generated /etc/fstab file on the system.

If you want to keep identifying the partitions using block device names, you can use the "--bootloader no-uuid" and the "--fstab no-uuid" options of g5k-postinstall, in the postinstalls/script of the description of your environment. Please refer to the "Customizing the postinstalls" section of the "Advanced Kadeploy" page: Advanced_Kadeploy#Using_g5k-postinstall

As an additional change, Kadeploy now tries to use kexec more often, which should make the first deployment of a job noticeably faster.

-- Grid'5000 Team 17:30, June 9th 2021 (CEST)


Read more news

Grid'5000 sites

Current funding

As from June 2008, Inria is the main contributor to Grid'5000 funding.

INRIA

Logo INRIA.gif

CNRS

CNRS-filaire-Quadri.png

Universities

Université Grenoble Alpes, Grenoble INP
Université Rennes 1, Rennes
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
Université Bordeaux 1, Bordeaux
Université Lille 1, Lille
École Normale Supérieure, Lyon

Regional councils

Aquitaine
Auvergne-Rhône-Alpes
Bretagne
Champagne-Ardenne
Provence Alpes Côte d'Azur
Hauts de France
Lorraine