Difference between revisions of "Grid5000:Home"

From Grid5000
Jump to: navigation, search
(Fix news link)
 
(297 intermediate revisions by 16 users not shown)
Line 1: Line 1:
 
__NOTOC__ __NOEDITSECTION__
 
__NOTOC__ __NOEDITSECTION__
 
 
{|width="95%"
 
{|width="95%"
 
|- valign="top"
 
|- valign="top"
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
[[Image:Logo_sc05.gif|64px|left|SC2005]]
+
[[Image:renater5-g5k.jpg|thumbnail|250px|right|Grid'5000]]
==Breaking news==
+
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and AI.'''
* '''Grid'5000 @ SC2005''': Grid'5000 platform demonstations are presented at SuperComputing (see [[Grid5000:News|here]]).
+
 
 +
Key features:
 +
* provides '''access to a large amount of resources''': 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
 +
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
 +
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
 +
* '''designed to support Open Science and reproducible research''', with full traceability of infrastructure and software changes on the testbed
 +
* '''a vibrant community''' of 500+ users supported by a solid technical team
 +
 
 +
<br>
 +
Read more about our [[Team|teams]], our [[Publications|publications]], and the [[Grid5000:UsagePolicy|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].
 +
 
 +
<b>Grid'5000 is merging with [https://fit-equipex.fr FIT] to build the [http://www.silecs.net/ SILECS Infrastructure for Large-scale Experimental Computer Science]. Read [http://www.silecs.net/wp-content/uploads/2018/04/Desprez-SILECS.pdf an Introduction to SILECS] (April 2018)</b>
 +
 
 +
<br>
 +
Recently published documents and presentations:
 +
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
 +
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]
 +
 
 +
Older documents:
 +
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
 +
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]
 +
 
 +
<br>
 +
Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL [[Hemera|HEMERA]] (2010-2014).
 
|}
 
|}
  
==Grid'5000 at a glance==
+
<br>
[[Image:CarteSites.JPG|thumbnail|128px|right|Grid'5000 sites]]
+
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
* '''Grid'5000''' project aims at building an '''experimental Grid platform''' gathering '''9 sites''' geographically distributed in France:
+
<br>
===Sites:===
+
 
{|width="75%" cellspacing="3"  
+
== Random pick of publications ==
 +
{{#publications:}}
 +
 
 +
==Latest news==
 +
<rss max=4 item-max-length="2000">https://www.grid5000.fr/w?title=News&action=feed&feed=atom</rss>
 +
----
 +
[[News|Read more news]]
 +
 
 +
=== Grid'5000 sites===
 +
{|width="100%" cellspacing="3"  
 
|- valign="top"
 
|- valign="top"
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Bordeaux
+
* [[Grenoble:Home|Grenoble]]
* Grenoble
+
* [[Lille:Home|Lille]]
* Lille
+
* [[Luxembourg:Home|Luxembourg]]
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Lyon
+
* [[Lyon:Home|Lyon]]
* Nancy
+
* [[Nancy:Home|Nancy]]
* Orsay
+
* [[Nantes:Home|Nantes]]
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
* Rennes
+
* [[Rennes:Home|Rennes]]
* Sophia-Antipolis
+
* [[Sophia:Home|Sophia-Antipolis]]
* Toulouse
+
* [[Toulouse:Home|Toulouse]]
 
|-
 
|-
 
|}
 
|}
  
* The main purpose of this platform is to serve as an experimental testbed for research in Grid Computing.
+
== Current funding ==
* This project is one initiative of the '''French ACI Grid''' Incentive [[Image:LogoACIGRID.jpg]]
+
As from June 2008, Inria is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].  
 
 
==Introduction==
 
* '''Grid'5000''' is a research effort developping a '''large scale nation wide infrastructure for Grid research'''.
 
 
 
* '''17 [[Grid5000:Laboratories|laboratories]]''' are involved, nation wide, in the objective of providing the community of Grid researchers a testbed allowing experiments in all the software layers between the network protocols up to the applications:
 
{|
 
|
 
{| border=1
 
|- bgcolor="#EEEEEE"
 
| Applications
 
|- bgcolor="#DDDDDD"
 
| Algorithms
 
|- bgcolor="#CCCCCC"
 
| Runtime
 
|- bgcolor="#BBBBBB"
 
| Middleware
 
|- bgcolor="#AAAAAA"
 
| Operating Systems
 
|- bgcolor="#999999"
 
| Network protocols
 
|}
 
| Grid'5000 will allow Grid experiments France wide in all these software layers
 
|}
 
The current plans are to assemble a physical platform featuring 9 clusters, each with 100 to a thousand PCs, connected by the [http://www.renater.fr RENATER] Education and Research Network. All clusters will be connected to Renater at 1 Gb/s (10 Gb/s is expected in near future).
 
 
 
This high collaborative research effort is funded by the French ministry of Education and Research, INRIA, CNRS, the Universities of all sites and some regional councils.
 
 
 
==Funding Institutions==
 
 
{|width="100%" cellspacing="3"
 
{|width="100%" cellspacing="3"
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Ministère de l'Education, de la Jeunesse et de la Recherche===
 
[[Image:Logo-Ministere-Recherche.jpg]]
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===ACI Grid===
 
[[Image:LogoACIGRID.jpg]]
 
 
|-
 
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===INRIA===
 
===INRIA===
[[Image:Logo-inria.gif]]
+
[[Image:Logo_INRIA.gif|300px]]
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===CNRS===
 
===CNRS===
[[Image:Logo-cnrs.gif]]
+
[[Image:CNRS-filaire-Quadri.png|125px]]
 
|-
 
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Universities===
 
===Universities===
University of Paris Sud, Orsay<br/>
+
Université Grenoble Alpes, Grenoble INP<br/>
University Joseph Fourier, Grenoble<br/>
+
Université Rennes 1, Rennes<br/>
University of Nice-Sophia Antipolis, Sophia Antipolis<br/>
+
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
University of Rennes 1, Rennes<br/>
+
Université Bordeaux 1, Bordeaux<br/>
University Paul Sabatier / FERIA, Toulouse<br/>
+
Université Lille 1, Lille<br/>
University Bordeaux 1, Bordeaux<br/>
+
École Normale Supérieure, Lyon<br/>
University Lille 1 / GENOPOLE, Lille<br/>
 
Ecole Normale Supérieure / MYRICOM, Lyon<br/>
 
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Regional councils===
 
===Regional councils===
 +
Aquitaine<br/>
 +
Auvergne-Rhône-Alpes<br/>
 
Bretagne<br/>
 
Bretagne<br/>
 +
Champagne-Ardenne<br/>
 
Provence Alpes Côte d'Azur<br/>
 
Provence Alpes Côte d'Azur<br/>
Aquitaine<br/>
+
Hauts de France<br/>
Ile de France<br/>
+
Lorraine<br/>
===General Councils===
 
Alpes Maritimes
 
 
|}
 
|}

Latest revision as of 14:30, 12 February 2021

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and AI.

Key features:

  • provides access to a large amount of resources: 15000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: PMEM, GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.

Grid'5000 is merging with FIT to build the SILECS Infrastructure for Large-scale Experimental Computer Science. Read an Introduction to SILECS (April 2018)


Recently published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2021-06-20 15:56): 1 current events, None planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2181 overall):

  • Mathias Fleury, Hans-Jörg Schurr. Reconstructing veriT Proofs in Isabelle/HOL. PxTP 2019 - Sixth Workshop on Proof eXchange for Theorem Proving, Aug 2019, Natal, Brazil. pp.36-50, 10.4204/EPTCS.301.6. hal-02276530 view on HAL pdf
  • Orcun Yildiz, Shadi Ibrahim. Preserving Fairness in Shared Hadoop Cluster: A Study on the Impact of (Non-) Preemptive Approaches. Research Report RR-9384, Inria Rennes - Bretagne Atlantique. 2020. hal-03091371 view on HAL pdf
  • Amir Wonjiga, Sean Peisert, Louis Rilling, Christine Morin. Blockchain as a Trusted Component in Cloud SLA Verification. UCC 2019 - 12th IEEE/ACM International Conference on Utility and Cloud Computing, Dec 2019, Auckland, New Zealand. pp.93-100, 10.1145/3368235.3368872. hal-02354647v3 view on HAL pdf
  • Remous-Aris Koutsiamanis, Georgios Papadopoulos, Bruno Quoitin, Nicolas Montavont. A Centralized Controller for Reliable and Available Wireless Schedules in Industrial Networks. MSN 2020 : 16th International Conference on Mobility, Sensing and Networking, Dec 2020, Virtual, Japan. pp.1-9, 10.1109/MSN50589.2020.00018. hal-02987638 view on HAL pdf
  • Nicolas Furnon, Romain Serizel, Irina Illina, Slim Essid. DNN-Based Distributed Multichannel Mask Estimation for Speech Enhancement in Microphone Arrays. ICASSP 2020 - 45th International Conference on Acoustics, Speech, and Signal Processing, May 2020, Barcelona, Spain. hal-02389159v3 view on HAL pdf


Latest news

Rss.svgDebian 11 "Bullseye" preview environments are now available for deployments

Debian 11 stable (Bullseye) will be released in a few weeks. We are pleased to offer a "preview" of kadeploy environments for Debian 11 (currently still Debian testing), that you can deploy already. See the debian11 environments in kaenv3.

New features and changes in Debian 11 are described in: https://www.debian.org/releases/bullseye/amd64/release-notes/ch-whats-new.en.html .

In particular, it includes many software updates:

  • Cuda 11.2 / Nvidia drivers 460.73.01
  • OpenJDK 17
  • Python 3.9.2
    • python3-numpy 1.19.5
    • python3-scipy 1.6.0
    • python3-pandas 1.1.5

(Note that Python 2 is not included as deprecated since Jan. 2020 and /usr/bin/python is symlinked to /usr/bin/python3)

  • Perl 5.32.1
  • GCC 9.3 and 10.2
  • G++ 10.2
  • Libboost 1.74.0
  • Ruby 2.7.3
  • CMake 3.18.4
  • GFortran 10.2.1
  • Liblapack 3.9.0
  • libatlas 3.10.3
  • RDMA 33.1
  • OpenMPI 4.1.0

Known regressions and problems are :

  • The std environment is not ready yet, and will not be the default Grid'5000 environment until official Debian 11 Bullseye release.
  • Cuda/Nvidia drivers do not support - out-of-the-box or at all - some quite old GPUs.
  • BeeGFS is not operational at the moment.

Let us know if you want us to support some tools, softwares,… that are not available on big images.

As a reminder, you can use the following commands to deploy an environment on nodes (https://www.grid5000.fr/w/Getting_Started#Deploying_nodes_with_Kadeploy):

 $ oarsub -t deploy -I
 $ kadeploy3 -e debian11-x64-big...

Rss.svgKadeploy: use of UUID partition identifiers and faster deployments

Up to now, kadeploy was identifying disk partitions with their block device names (e.g. /dev/sda3) when deploying a system. This no longer works reliably because of disk inversion issues with recent kernels. As a result, we have changed kadeploy to use filesystem UUIDs instead.

This change affects the root partition passed to the kernel command line as well as the generated /etc/fstab file on the system.

If you want to keep identifying the partitions using block device names, you can use the "--bootloader no-uuid" and the "--fstab no-uuid" options of g5k-postinstall, in the postinstalls/script of the description of your environment. Please refer to the "Customizing the postinstalls" section of the "Advanced Kadeploy" page: Advanced_Kadeploy#Using_g5k-postinstall

As an additional change, Kadeploy now tries to use kexec more often, which should make the first deployment of a job noticeably faster.

-- Grid'5000 Team 17:30, June 9th 2021 (CEST)

Rss.svgNew monitoring service with Kwollect is now stable

The new Grid'5000 monitoring service based on Kwollect is now stable. Kwollect will now serve requests adressed to Grid'5000 "Metrology API", i.e., from this URL:

https://api.grid5000.fr/stable/sites/SITE/metrics

The former API based on Ganglia is no longer available, and Ganglia will be removed from Grid'5000 environments starting from the next Debian version (debian11).

Main features of Kwollect are:

  • Focus on "environmental" monitoring, i.e. metrics not available from inside the nodes: electrical consumption, temperature, metrics from network equipments or nodes' BMC… but Kwollect also monitors node's metrics from Prometheus exporters
  • Support for Grid'5000 wattmeters at high frequency
  • On-demand activation of optional metrics
  • Custom metrics can be pushed by users
  • Grafana-based vizualisation

Its usage of this new service is described at: Monitoring_Using_Kwollect

Here are the main other changes since the last annoucement on December:

  • Monitoring related entries in Reference API have been cleaned (nodes' "sensors" and "wattmetre" keys removed, "pdu" moved at top level), OAR property wattmeter=SHARED and wattmeter=MULTIPLE removed
  • Visualization now uses Grafana
  • Metrics naming and polling period have been updated
  • Grid'5000 documentation has been updated
  • Bugs and performance fixes

-- Grid'5000 Team 12:00, May 21th 2021 (CEST)

Rss.svgUpgrade of the graffiti-13 node in Nancy

We are happy to announce that the graffiti-13 node on the Nancy site in production queue has been upgraded:

its 4 GeForce RTX 2080 Ti 11GB have been replaced by 4 RTX 6000 24GB GDDR6[1].

Reservation example : oarsub -q production -p "host='graffiti-13.nancy.grid5000.fr'" -I

[1] https://www.nvidia.com/fr-fr/design-visualization/quadro/rtx-6000/

-- Grid'5000 Team 17:00, May 6th 2021 (CEST)


Read more news

Grid'5000 sites

Current funding

As from June 2008, Inria is the main contributor to Grid'5000 funding.

INRIA

Logo INRIA.gif

CNRS

CNRS-filaire-Quadri.png

Universities

Université Grenoble Alpes, Grenoble INP
Université Rennes 1, Rennes
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
Université Bordeaux 1, Bordeaux
Université Lille 1, Lille
École Normale Supérieure, Lyon

Regional councils

Aquitaine
Auvergne-Rhône-Alpes
Bretagne
Champagne-Ardenne
Provence Alpes Côte d'Azur
Hauts de France
Lorraine