Skip to content
GitHub

About P1 Compute Resources

The Pioneer Centre for Artificial Intelligence (P1) is a Danish cross-university research centre and consortium dedicated to advancing AI research through collaboration with partnering institutions.

This documentation aims to centralize practical knowledge and provide a concise guide for P1 affiliates on how to access and use our available high-performance computing (HPC) facilities.

Alt Text

A central mission of P1 is to foster broad research collaboration, both within the centre and across its affiliated universities. Our HPC resources are designed to support P1 researchers at partner institutions, enabling projects of varying scale and complexity.

For more information, visit the P1 Homepage or P1 Computing.

While we prioritize smaller projects as they represent the majority of research needs, we also support larger projects through our tiered compute resources, including the capability to train very large models and support European-level workloads on Tier 2/1 facilities. But, very large scale projects might be better supported within their own funding.

In practical terms this means that while training a large language or vision model from scratch might not be suitable for our Tier 3 facilities, such projects can be accommodated on a Tier 2 national facility (ex. Gefion HPC), which is specifically designed for large-scale model training and can even support some European-level (Tier 1) workloads.

In general the P1 does not have official support staff. Instead we rely on the community to help each other. However, for questions regarding the cluster, we have the following two contact e-mails:

If your submission to join the cluster has been pending for more than a week, you can contact the member of the P1 HPC gouvernance board of your institution. General announcements regarding the cluster will be made in the #compute channel on the official P1-slack. Similarly if you have general questions regarding compute, including using the P1-cluster, feel free to post those questions in the #ask-compute channel.

HPC Gouvernance Board

Nicki Skafte Detlefsen

DTU

Technical University of Denmark

nsde@dtu.dk

Rob van der Goot

ITU

IT University of Copenhagen

robv@itu.dk

Mikkel Fruelund Odgaard

KU

University of Copenhagen

miod@di.ku.dk

Generally, P1 can help purchasing hardware for you if it is to be a part of P1’s clusters. In this setup you would have priority access to the hardware that you purchase, but if it is unused other researchers at P1 can use it as well. Most HPC hardware has an expected utilization time of approximately 5 years.

An example of a compute budget expenditure:

A researcher might have 400k DKK for non-GDPR computing. In that case P1 can purchase a 2xA100 node, with the researcher getting priority access. With the last 100k DKK, we recommend a diversification of the researchers compute portfolio.

Using some of the 100k for personal computers with powerful GPU’s and reserving a significant portion of funds for later. This could be for storage or cloud compute. Having the budget to spring for cloud compute if peak pressure on other computing resources makes training a model in time for a submission deadline infeasible can be necessary.

Depending on what you are doing, you will need different levels of compute. The following table gives an overview of the different levels of compute and what they are suitable for.

TierOrganization levelDescription
4PersonalPersonal laptop and/or desktop with either no GPUs or a single one. At this level compute is limited and mostly consists of code development and running experiments with small datasets and models.
3InstituteThe next level is compute clusters containing either several smaller GPUs or a few larger ones. Multiple experiments can be run in parallel and models can be parallelized over a few GPUs, but it is not feasible to train very large models from scratch. Expect limitations to wall time (a couple of days), storage (around ½-1 TB) and number of nodes/GPUs used (around 2-16).
2NationalNational HPC facilities provide the necessary resources for large-scale model training. Expect fewer restrictions to wall time (1-2 weeks), storage (around 1-10 TB) and number of nodes/GPUs used (around 16-64). Some national facilities can also support European-level (Tier 1) workloads. For example, P1’s Gefion HPC is a national facility that can support both Tier 2 and possibly Tier 1 workloads.
1EuropeanEuropean HPC centers offer significantly larger scale compute resources, typically providing easy access to 8-16 GPUs with the possibility of scaling up to 500-800 GPUs depending on project requirements and the specific HPC center. Unlike lower tiers, walltime is usually not restricted; instead, limitations are set on total GPU hours used. See Karolina and Meluxina for examples of European HPC centers with GPUs.
0GlobalThe highest level of compute. Possibility to use above 1000 GPUs and perform calculations at exascale. Usually only available to large research projects and collaborations with industry. See LUMI for an example of a global HPC center with GPUs.

The P1 DTU and NGC HPCs sit at level 3 in the hierarchy, while Gefion HPC primarily operates at level 2 but can support some level 1 workloads, making it capable of supporting both national and some European-level research projects.