Difference between revisions of "Grid5000:Home"

From Grid5000
Jump to: navigation, search
(46 intermediate revisions by 6 users not shown)
Line 4: Line 4:
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|bgcolor="#f5fff5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
[[Image:renater5-g5k.jpg|thumbnail|250px|right|Grid'5000]]
 
[[Image:renater5-g5k.jpg|thumbnail|250px|right|Grid'5000]]
 
+
'''Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and IA.'''
'''Grid'5000 is a large-scale and versatile testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data.'''
 
  
 
Key features:
 
Key features:
* provides '''access to a large amount of resources''': 1000 nodes, 8000 cores, grouped in homogeneous clusters, and featuring various technologies: 10G Ethernet, Infiniband, GPUs, Xeon PHI
+
* provides '''access to a large amount of resources''': 12000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
 
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
 
* '''highly reconfigurable and controllable''': researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
 
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
 
* '''advanced monitoring and measurement features for traces collection of networking and power consumption''', providing a deep understanding of experiments
Line 15: Line 14:
  
 
<br>
 
<br>
Read more about our [[Grid5000:People|teams]], our [[Grid5000:Publications|publications]], and the [[Grid5000:UserCharter|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].
+
Read more about our [[Team|teams]], our [[Publications|publications]], and the [[Grid5000:UsagePolicy|usage policy]] of the testbed. Then [[Grid5000:Get_an_account|get an account]], and learn how to use the testbed with our [[Getting_Started|Getting Started tutorial]] and the rest of our [[:Category:Portal:User|Users portal]].
<!--
+
 
 +
<b>Grid'5000 is merging with [https://fit-equipex.fr FIT] to build the [http://www.silecs.net/ SILECS Infrastructure for Large-scale Experimental Computer Science]. Read [http://www.silecs.net/wp-content/uploads/2018/04/Desprez-SILECS.pdf an Introduction to SILECS] (April 2018)</b>
 +
 
 
<br>
 
<br>
Recently published documents:
+
Recently published documents and presentations:
* Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014 (FIXME)
+
* [[Media:Grid5000.pdf|Presentation of Grid'5000]] (April 2019)
* Report from the Grid'5000 Science Advisory Board meeting (2014) (FIXME)
+
* [https://www.grid5000.fr/mediawiki/images/Grid5000_science-advisory-board_report_2018.pdf Report from the Grid'5000 Science Advisory Board (2018)]
-->
+
 
 +
Older documents:
 +
* [https://www.grid5000.fr/slides/2014-09-24-Cluster2014-KeynoteFD-v2.pdf Slides from Frederic Desprez's keynote at IEEE CLUSTER 2014]
 +
* [https://www.grid5000.fr/ScientificCommittee/SAB%20report%20final%20short.pdf Report from the Grid'5000 Science Advisory Board (2014)]
  
 
<br>
 
<br>
Line 27: Line 31:
 
|}
 
|}
  
== Latest publications from Grid'5000 users ==
 
{{#publications:3||*}}
 
 
<br>
 
<br>
 
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
 
{{#status:0|0|0|http://bugzilla.grid5000.fr/status/upcoming.json}}
==Latest news==
+
<br>
=== [http://www.lifl.fr/~melab/HTML/Journee-G5K-Lille.htm Grid'5000 tutorial days in Lille] ===
 
We are happy to let you known that tutorials around Grid'5000 will be organized in Lille on November 20th, 2014, with a few seats available for people outside Lille. All information on the [http://www.lifl.fr/~melab/HTML/Journee-G5K-Lille.htm dedicated web page].
 
=== [[Grid5000:School2014|Grid'5000 spring school]] now finished ===
 
The Grid'5000 spring school took place between June 16th, 2014 and June 19th, 2014 in Lyon. Three awards were given for presentation or challenge entries (the challenge entries ended as a tie):
 
{|width="75%" cellspacing="3"
 
|- valign="top"
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
[[Image:Presentation_Gistau_award.png|252px|Best presentation award to Miguel Liroz Gistau]]
 
  
Best presentation award to Miguel Liroz Gistau, Reza Akbarinia and Patrick Valduriez
+
== Random pick of publications ==
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
+
{{#publications:}}
[[Image:Challenge_Buchert_award.png|252px|Best challenge entry to Tomasz Buchert, Emmanuel Jeanvoine and Lucas Nussbaum]]
 
  
Best challenge entry to Tomasz Buchert, Emmanuel Jeanvoine and Lucas Nussbaum
+
==Latest news==
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
+
<rss max=4 item-max-length="2000">https://www.grid5000.fr/mediawiki/index.php?title=News&action=feed&feed=atom</rss>
[[Image:Challenge_Pastor_award.png|252px|Best challenge entry to Jonathan Pastor and Laurent Pouilloux]]
 
 
 
Best challenge entry to Jonathan Pastor and Laurent Pouilloux
 
|-
 
|}
 
<!--
 
=== Journée Aramis ===
 
 
 
Le 16/04 aura lieu la journée Aramis "Usages et services du cloud dans notre environnement enseignement supérieur et recherche" avec une présentation de Grid'5000 par Simon Delamare.
 
 
 
Cette journée est webcastée par la cellule Webcast du CC-IN2P3. Vous pouvez donc suivre en direct les [http://webcast.in2p3.fr/live/pleiniere_aramis_2014 présentations]. Le programme de la journée est disponible [http://aramis.resinfo.org/wiki/doku.php?id=pleniaires:pleniere17avril2014 ici]
 
 
 
=== [[Grid5000:UserCharter|Grid'5000 Charter]] revised ===
 
The Grid'5000 charter has been reworded and simplified so as to be more inclusive about possible uses. The rules to ensure availability during the day for any user and during the night for large resource usage have not been changed.
 
=== [[Grid5000:school2014|Grid'5000 school 2014]] announced ===
 
Organized in Lyon, from June 16th to June 19th 2014, this [[Grid5000:school2014|spring school]] will bring together, but is not limited to, Grid'5000's users, technical team and executive committee for 4 days of tutorials and talks focusing on best-practices and results. Presentations and practical sessions will cover both basic usage of the platform, for new users, or potential users of Grid'5000 and advanced and new usage of the platform, for current users. Deadline to submit presentation proposals is March 29th, 2014.
 
 
 
=== Grid'5000 users win second prize at CCGRID 2013's SCALE challenge ===
 
Snooze based entry running on Grid'5000 entry wins [http://www.pds.ewi.tudelft.nl/ccgrid2013/awards/ 2nd prize] at [http://www.pds.ewi.tudelft.nl/ccgrid2013/ CCGrid 2013] [http://www.pds.ewi.tudelft.nl/ccgrid2013/calls/scale-challenge/ SCALE challenge]: well done Matthieu and Anne-Cécile for defending the entry titled ''Scalability of the Snooze Autonomic Cloud Management System'' by Eugen Feller, Christine Morin, Matthieu Simonin, Anne-Cécile Orgerie, and Yvon Jégou.
 
 
 
=== Grid'5000 users finalists of the SCALE'2013 challenge ===
 
Two submissions (out of five) from Grid'5000 users took part in the final of the international SCALE'2013 challenge (held with CCGrid'2013):
 
* D. Balouek, A. Lèbre, F. Quesnel Flauncher and DVMS -- Deploying and Scheduling Thousands of Virtual Machines on Hundreds of Nodes Distributed Geographically
 
* Eugen Feller, Christine Morin, Matthieu Simonin, Anne-Cécile Orgerie, and Yvon Jégou -- Scalability of the Snooze Autonomic Cloud Management System.
 
 
 
The first proposition presents the deployment and scheduling of thousands of virtual machines, conducted with the Flauncher and DVMS frameworks, across the Grid'5000 testbed.
 
The frameworks have been able to deploy and schedule up to 10000 VMs during the tests. This research has been conducted in the context of the INRIA Hemera initiative.
 
 
 
The second proposition presents the Snooze architecture and focus on the following aspects :
 
* System set up scalability and resources consumption,
 
* Self-healing capabilities.
 
The deployment of Snooze were distributed over several sites of the Grid'5000 testbed.
 
The framework has been able to start 11000 of system services, recover thousands of failures
 
and used to launch large hadoop/mapreduce experiments.
 
This research has been conducted in the context of the INRIA Snooze ADT.
 
=== Grid'5000 tutorial during ComPAS'2013 ===
 
The [http://compas2013.inrialpes.fr/ ComPAS'2013 conference] (replacing RenPar, SympA and CFSE), to be held in Grenoble between January 15th and 18th, will feature a Grid'5000 tutorial.
 
-->
 
 
----
 
----
[[Grid5000:News|read more news]]
+
[[News|Read more news]]
  
 
=== Grid'5000 sites===
 
=== Grid'5000 sites===
{|width="75%" cellspacing="3"  
+
{|width="100%" cellspacing="3"  
 
|- valign="top"
 
|- valign="top"
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
Line 95: Line 50:
 
* [[Lille:Home|Lille]]
 
* [[Lille:Home|Lille]]
 
* [[Luxembourg:Home|Luxembourg]]
 
* [[Luxembourg:Home|Luxembourg]]
 +
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
* [[Lyon:Home|Lyon]]
 
* [[Lyon:Home|Lyon]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
 
* [[Nancy:Home|Nancy]]
 
* [[Nancy:Home|Nancy]]
 
* [[Nantes:Home|Nantes]]
 
* [[Nantes:Home|Nantes]]
* [[Reims:Home|Reims]]
+
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
* [[Rennes:Home|Rennes]]
 
* [[Rennes:Home|Rennes]]
|width="33%" bgcolor="#f5f5f5" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
 
* [[Sophia:Home|Sophia-Antipolis]]
 
* [[Sophia:Home|Sophia-Antipolis]]
 
* [[Toulouse:Home|Toulouse]]
 
* [[Toulouse:Home|Toulouse]]
Line 108: Line 62:
  
 
== Current funding ==
 
== Current funding ==
As from June 2008, INRIA is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].  
+
As from June 2008, Inria is the main contributor to [[Grid5000:Funding|Grid'5000 funding]].  
 
{|width="100%" cellspacing="3"
 
{|width="100%" cellspacing="3"
 
|-
 
|-
Line 116: Line 70:
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===CNRS===
 
===CNRS===
[[Image:CNRS-filaire-MonoBleu.gif|100px]]
+
[[Image:CNRS-filaire-Quadri.png|125px]]
 
|-
 
|-
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Universities===
 
===Universities===
University Joseph Fourier, Grenoble<br/>
+
Université Grenoble Alpes, Grenoble INP<br/>
University of Rennes 1, Rennes<br/>
+
Université Rennes 1, Rennes<br/>
 
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
 
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse<br/>
University Bordeaux 1, Bordeaux<br/>
+
Université Bordeaux 1, Bordeaux<br/>
University Lille 1, Lille<br/>
+
Université Lille 1, Lille<br/>
Ecole Normale Supérieure, Lyon<br/>
+
École Normale Supérieure, Lyon<br/>
Université de Reims Champagne-Ardenne, Reims<br/>
 
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
| width="50%" bgcolor="#f5f5f5" valign="top" align="center" style="border:1px solid #cccccc;padding:1em;padding-top:0.5em;"|
 
===Regional councils===
 
===Regional councils===
 
Aquitaine<br/>
 
Aquitaine<br/>
 +
Auvergne-Rhône-Alpes<br/>
 
Bretagne<br/>
 
Bretagne<br/>
 
Champagne-Ardenne<br/>
 
Champagne-Ardenne<br/>
 
Provence Alpes Côte d'Azur<br/>
 
Provence Alpes Côte d'Azur<br/>
Nord Pas de Calais<br/>
+
Hauts de France<br/>
 
Lorraine<br/>
 
Lorraine<br/>
 
|}
 
|}

Revision as of 07:23, 4 October 2019

Grid'5000

Grid'5000 is a large-scale and flexible testbed for experiment-driven research in all areas of computer science, with a focus on parallel and distributed computing including Cloud, HPC and Big Data and IA.

Key features:

  • provides access to a large amount of resources: 12000 cores, 800 compute-nodes grouped in homogeneous clusters, and featuring various technologies: GPU, SSD, NVMe, 10G and 25G Ethernet, Infiniband, Omni-Path
  • highly reconfigurable and controllable: researchers can experiment with a fully customized software stack thanks to bare-metal deployment features, and can isolate their experiment at the networking layer
  • advanced monitoring and measurement features for traces collection of networking and power consumption, providing a deep understanding of experiments
  • designed to support Open Science and reproducible research, with full traceability of infrastructure and software changes on the testbed
  • a vibrant community of 500+ users supported by a solid technical team


Read more about our teams, our publications, and the usage policy of the testbed. Then get an account, and learn how to use the testbed with our Getting Started tutorial and the rest of our Users portal.

Grid'5000 is merging with FIT to build the SILECS Infrastructure for Large-scale Experimental Computer Science. Read an Introduction to SILECS (April 2018)


Recently published documents and presentations:

Older documents:


Grid'5000 is supported by a scientific interest group (GIS) hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations. Inria has been supporting Grid'5000 through ADT ALADDIN-G5K (2007-2013), ADT LAPLACE (2014-2016), and IPL HEMERA (2010-2014).


Current status (at 2020-02-19 16:05): No current events, None planned (details)


Random pick of publications

Five random publications that benefited from Grid'5000 (at least 2139 overall):

  • Dimitri Delabroye, Simon Delamare, David Loup, Lucas Nussbaum. Remplacer un routeur par un serveur Linux : retour d'expérience des passerelles d'accès à Grid'5000. JRES - Journées Réseaux de l'Enseignement et de la Recherche, Dec 2019, Dijon, France. hal-02401684 view on HAL pdf
  • Abdulqawi Saif, Lucas Nussbaum, Ye-Qiong Song. On the Impact of I/O Access Patterns on SSD Storage. Research Report RR-9319, Inria. 2020. hal-02430564 view on HAL pdf
  • Abdulqawi Saif, Alexandre Merlin, Olivier Dautricourt, Maël Houbre, Lucas Nussbaum, et al.. Emulation of Storage Performance in Testbed Experiments with Distem. CNERT 2019 - IEEE INFOCOM International Workshop on Computer and Networking Experimental Research using Testbeds, Apr 2019, Paris, France. pp.6. hal-02078301 view on HAL pdf
  • Francieli Zanon Boito, Jean-François Méhaut, Thierry Deutsch, Brice Videau, Frédéric Desprez. Instrumental Data Management and Scientific Workflow Execution: the CEA case study. IPDPSW 2019 - International Parallel and Distributed Processing Symposium Workshops, May 2019, Rio de Janeiro, Brazil. pp.850-857, 10.1109/IPDPSW.2019.00139. hal-02076963 view on HAL pdf
  • Jean Luca Bez, Francieli Zanon Boito, Ramon Nou, Alberto Miranda, Toni Cortes, et al.. Detecting I/O Access Patterns of HPC Workloads at Runtime. SBAC-PAD 2019 - International Symposium on Computer Architecture and High Performance Computing, Oct 2019, Campo Grande, Brazil. hal-02276191 view on HAL pdf


Latest news

Rss.svgNew cluster "troll" available in Grenoble

We have the pleasure to announce that a new cluster called "troll" is available in Grenoble¹.

It features 4 Dell R640 nodes with 2 Intel® Xeon® Gold 5218, 16 cores/CPU, 384GB DDR4, 1.5 TB PMEM (Intel® Optane™ DC Persistent Memory)²³, 1.6 TB NVME SSD, 10Gbps Ethernet, and 100Gb Omni-Path.

Energy monitoring⁴ is available for this cluster, provided by the same devices used for the other clusters in Grenoble.

This cluster has been funded by the PERM@RAM project from Laboratoire d'Informatique de Grenoble (CNRS/INS2I grant).


¹: https://www.grid5000.fr/w/Grenoble:Hardware

²: https://software.intel.com/en-us/articles/quick-start-guide-configure-intel-optane-dc-persistent-memory-on-linux

³: https://docs.pmem.io/persistent-memory/

⁴: https://www.grid5000.fr/w/Energy_consumption_monitoring_tutorial

-- Grid'5000 Team 17:00, February 3rd 2020 (CET)


Rss.svgNew cluster available in Nancy: grue (20 GPUs)

We have the pleasure to announce that the Grue cluster in Nancy¹ (production queue) is now available:

It features 5 Dell R7425 servers nodes with four Tesla T4², 128 GB DDR4, 1x480 GB SSD, 2 x AMD EPYC 7351, 16 cores/CPU

As this cluster features 4 GPU per node, we remind you that you can monitor GPU (and node) usage using the Ganglia tool (std environment only):

If your experiments do not require all the GPU of a single node, it is possible to reserve GPU³ at the resource level (see https://grid5000.fr/w/News#Enabling_GPU_level_resource_reservation_in_OAR for some examples). You can also use the nvidia-smi and htop commands on your reserved nodes to get more information about your GPU/CPU usage.

This cluster has been funded by Ihe CPER LCHN project (Langues, Connaissances & Humanités Numériques, Contrat de plan État / Région Lorraine 2015-2020), and the LARSEN and MULTISPEECH teams at LORIA / Inria Nancy Grand Est.

As a reminder, since this cluster is part of the "production" queue, specific usage rules apply.


¹: https://www.grid5000.fr/w/Hardware

²: https://www.nvidia.com/content/dam/en-zz/Solutions/Data-Center/tesla-t4/t4-tensor-core-d...

Rss.svgGrid'5000 users survey

We are conducting a survey to help us better understand your needs and make Grid'5000 a better research infrastructure.

We thank you in advance for taking a few minutes to complete it (you can answer in French if you prefer).

The survey is available at:

https://sondages.inria.fr/index.php/672895

It will be open until December, 13rd.

-- Grid'5000 Team 15:00, November 26th 2019 (CET)

Rss.svgNew cluster "gemini" available at Lyon

We have the pleasure to announce you the availability of the new cluster "gemini" at Lyon.

Gemini includes two "Nvidia DGX-1" nodes, each with 8 Nvidia V100 GPUs, 2 Intel Xeon E5-2698 v4 @ 2.20GHz CPUs, 512GB DDR4, Infiniband EDR and 10Gbps Ethernet interfaces and 4 reservable¹ SSD disks.

Energy monitoring is also available for this cluster, provided by the same devices used for the other clusters in Lyon².

Remember that if you don't need the 8 GPUs, individual GPU may be reserved³. A script to install nvidia-docker is also available if you want to use Nvidia's images built for Docker⁴.

This cluster has been funded by the CPER LECO++ Project (FEDER, Région Auvergne-Rhone-Alpes, DRRT, Inria).

¹: https://www.grid5000.fr/w/Disk_reservation

²: https://www.grid5000.fr/w/Energy_consumption_monitoring_tutorial

³: https://www.grid5000.fr/w/Accelerators_on_Grid5000#Reserving_GPU_units_on_nodes_with_many_GPUs

⁴: https://www.grid5000.fr/w/Docker#Nvidia-docker

-- Grid'5000 Team 15:00, November 12th 2019 (CET)


Read more news

Grid'5000 sites

Current funding

As from June 2008, Inria is the main contributor to Grid'5000 funding.

INRIA

Logo INRIA.gif

CNRS

CNRS-filaire-Quadri.png

Universities

Université Grenoble Alpes, Grenoble INP
Université Rennes 1, Rennes
Institut National Polytechnique de Toulouse / INSA / FERIA / Université Paul Sabatier, Toulouse
Université Bordeaux 1, Bordeaux
Université Lille 1, Lille
École Normale Supérieure, Lyon

Regional councils

Aquitaine
Auvergne-Rhône-Alpes
Bretagne
Champagne-Ardenne
Provence Alpes Côte d'Azur
Hauts de France
Lorraine