P1 DTU HPC
The P1 DTU HPC is hosted at DTU and provides high-performance computing resources for P1 members (PhD and above). It is particularly suitable for medium to large-scale machine learning experiments and research projects.
Getting Access
Section titled “Getting Access”Requirements: PhD or higher (exceptions may apply), valid Danish university email, and registered P1 affiliation.
P1 Affiliation Form
P1 HPC Access Exception Form
DTU Account Signup Form
Connecting to the Cluster
Section titled “Connecting to the Cluster”The compute cluster is accessible at login9.hpc.dtu.dk
via SSH. Note that:
- Home directories have limited storage (30GB)
- Additional storage is available at
/dtu/p1/
- Interactive node is available for package installation and test runs
- Heavy jobs should be submitted as batch jobs
Access from Outside DTU Network
Section titled “Access from Outside DTU Network”- Download Cisco AnyConnect VPN client (see OpenConnect for Linux)
- Go to https://dtubasen.dtu.dk and sign in via Azure multi-factor auth using your full DTU username
- Set up multi-factor authentication
- Connect to vpn.dtu.dk using AnyConnect
- SSH to
login9.hpc.dtu.dk
using your DTU credentials
For persistent access, you can set up SSH keys:
-
Generate key
ssh-keygen -t ed25519 -f ~/.ssh/keyname -
Copy public key
ssh-copy-id -i ~/.ssh/keyname.pub username@login9.hpc.dtu.dk -
Connect
ssh -i ~/.ssh/keyname username@login9.hpc.dtu.dk
TIP: Consider setting up a SSH host alias for login9.hpc.dtu.dk
in your ~/.ssh/config
file to make it easier to connect to the cluster.
Support
Section titled “Support”Technical Support
Policy Support
General Questions
Compute Coordinator
For more technical information, refer to the P1 compute cluster documentation at DTU DCC.
Fair Use Policy
Section titled “Fair Use Policy”The following rules are in place to ensure fair use of the P1 DTU HPC:
- Maximum wall time: 72 hours
- Maximum number of GPUs in a job: 2 (one node)
- Maximum running jobs: ~50% of total available GPUs
- 500 gb of storage (+30gb in home directory)
If you have a project that requires more storage resources than the above, please contact the governance group compute-governance-p1@aicentre.dk to discuss your needs.
Specification
Section titled “Specification”- 7 Lenovo ThinkSystem SR665 V3 servers
- Each node specifications:
- 2 AMD EPYC 9354 32-Core Processors
- 768GB RAM
- 2 NVIDIA H100 PCIe GPUs (80GB each)
- Storage: 60TiB shared storage
- Operating System: Alma Linux
- Scheduling Environment: LSF
- Resource Allocation:
- 7 nodes available for batch jobs (queue:
p1
) - 1 node reserved for interactive usage (queue:
p1i
)
- 7 nodes available for batch jobs (queue: