Cluster Overview
The EMCC cluster provides high-performance computing (HPC) resources specifically tailored for Cryo-EM, structural biology, and computational chemistry research at Aarhus University.
It plays a central role in supporting the full Cryo-EM pipeline—from raw image pre-processing and motion correction, to particle picking, 3D reconstruction, and atomic model building using tools such as RELION, cryoSPARC, AlphaFold3, Rosetta, and Phenix.
EMCC includes a diverse array of CPU and GPU nodes optimized for both throughput and memory-intensive workloads. GPU nodes are configured for deep learning–based structure prediction and design tasks, while CPU nodes efficiently handle parallel reconstruction and refinement workflows.
Backed by a fault-tolerant 3.2 PB CephFS storage system with high-speed 100 GbE networking, EMCC enables efficient management of terabyte-scale Cryo-EM datasets and downstream analysis. All project data is regularly backed up to an off-site ZFS-based tier with snapshot recovery, ensuring data integrity and continuity.
Beyond Cryo-EM, EMCC supports protein crystallography, cryo-electron tomography, de novo protein design, molecular dynamics, and large language model applications—including retrieval-augmented generation (RAG) and scientific chatbots trained on in-house datasets.
Access to EMCC is available via secure SSH connections using AU’s jump-host infrastructure, the VDI.au.dk virtual desktop environment, or through AU's VPN service. Researchers can work interactively or submit batch jobs from remote terminals, workstations, or graphical desktop sessions.
In addition, EMCC hosts 13 high-end GPU-accelerated workstations in a dedicated on-site Graphics Room, allowing users to physically access powerful nodes for interactive 3D visualization, structural modeling, and real-time GPU workflows.
The cluster is available free of charge for pilot usage (up to 1 TB), with scalable project quotas (seats of 20 TB) and institutional support.
Research Fields & Associated Tools
Cryo-EM
Tomography
X-ray Crystallography
Protein Design
Large Language Models
Your Area Here?
Quick Acess Guide
Login via VDI (no VPN needed)
- If your a registered user go to vdi.au.dk.
-
Install the VMware Horizon Client when prompted, then launch it and add the server
vdi.au.dk
. -
Log in with your AU ID:
au123456@uni.au.dk
and your AU password.
You’ll be asked for 2FA (same as other AU services). - Pick any desktop that has “(public)” in its name.
- Inside the desktop, log in to EMCC with your EMCC username/password (not your AU password).
Direct SSH (alternative)
Terminal-only access using SSH or SFTP is also possible:
- If you’re off campus, connect to the AU VPN first.
-
SSH or SFTP into one of the public EMCC access hosts (the same ones marked “(public)” in VDI).
ssh/sftp emcc_username@<host-name>.bioxray.au.dk;
- You will be prompted for a password, which is your EMCC password, not your AU password.
Our Compute Nodes
CPU Nodes
Node | CPU Type | Cores | Threads | RAM Type | RAM (GB) | Network |
---|---|---|---|---|---|---|
vader1 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader2 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader3 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader4 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader5 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader6 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader7 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader8 | Xeon E5-2699 v3 | 36 | 72 | DDR4-2133 MHz | 384 | 100 GbE |
vader9 | AMD EPYC 9334 | 64 | 128 | DDR5-4800 MHz | 384 | 100 GbE |
vader10 | AMD EPYC 9334 | 64 | 128 | DDR5-4800 MHz | 384 | 100 GbE |
GPU Nodes
Node | CPU Type | Cores | Threads | RAM Type | RAM (GB) | Network | GPU Type |
---|---|---|---|---|---|---|---|
snoke4 | Xeon Gold 6226 | 32 | 64 | DDR4-2933 MHz | 384 | 100 GbE | 4× Tesla V100S |
snoke5 | Intel Platinum 8358 | 64 | 128 | DDR4-3200 MHz | 512 | 100 GbE | 4× Tesla A100 |
snoke6 | Intel Platinum 8358 | 64 | 128 | DDR4-3200 MHz | 1024 | 100 GbE | 4× Tesla A100 |
snoke7 | AMD EPYC 9334 | 64 | 128 | DDR5-4800 MHz | 768 | 100 GbE | 4× Tesla A100 |
snoke8 | AMD EPYC 9334 | 64 | 128 | DDR5-4800 MHz | 768 | 100 GbE | 4× Tesla L40S |
snoke9 | Intel Platinum 8562Y+ | 64 | 128 | DDR5-5600 MHz | 1024 | 100 GbE | 4× Tesla L40S |
snoke10 | Xeon Gold 6548N | 64 | 128 | DDR5-5600 MHz | 1024 | 100 GbE | 4× Tesla L40S |
Infrastructure
Primary Storage
Name: EMCC Storage (CephFS)
Capacity: 3.2 PB raw (16 nodes, 292 disks)
Snapshots: Nightly (7-day retention)
Mounts: /u/<user>
, /emcc/<project>
Access: CephFS, SSHFS
Data type: Non-sensitive research data (e.g., imaging, structural prediction)
Backup & Archiving
System: Remote ZFS pool (off-site, 1540-K13)
Backup: Nightly incremental (3-month retention)
Archive: Data optionally replicated to ERDA (30-day snapshots)
Data type: Non-sensitive research data (e.g., imaging, structural prediction)
Access & Security
Auth: IPA/IDM login (user/pass, SSH keys)
ACLs: POSIX permissions
Encryption: In transit only (Ceph msgr2, SSH/SFTP)
Compliance: GDPR & AU InfoSec Policy
Do not store sensitive or personal data.
Quotas & Costs
Default: 1 TB/user (free) Software‑stack documentation, step‑by‑step tutorials, general HPC info etc. NB. AU credentials required. Chat with our open‑source language models (Gemma, Qwen, DeepSeek ∞) or ask the RAG assistants for help with Data Management Plans, CryoSPARC and more — all served from EMCC hardware. NB. User account required! See the storage usage (Qutoa, % used, number of files etc) for each user/group account for both the EMCC cluster and the USERHOME. The stats are update each midnight NB. The site is only accessible from inside AU network! Click to email our help desk
Project: 20 TB/project (paid)
Annual Cost: 50.000 DKK/year (after pilot)
Scalability: Online expansion (Ceph + ZFS)
Documentation
LabBook Wiki
EMCC Chatbot
EMCC Diskhogs
Main Contact
EMCC Support – Main contact
Individual contacts
Rune Kidmose
rtk@mbg.au.dk
Special ConsultantJesper Lykkegaard Karlsen
jelka@au.dk
Chief ConsultantThomas Boesen
For EM-Facility related questions: thb@inano.au.dk
Cryo‑EM Manager