Welcome to the EMCC Cluster

High-performance computing for structural biology and life science research at Aarhus University.

Cluster Overview

The EMCC cluster provides high-performance computing (HPC) resources specifically tailored for Cryo-EM, structural biology, and computational chemistry research at Aarhus University.

It plays a central role in supporting the full Cryo-EM pipeline—from raw image pre-processing and motion correction, to particle picking, 3D reconstruction, and atomic model building using tools such as RELION, cryoSPARC, AlphaFold3, Rosetta, and Phenix.

EMCC includes a diverse array of CPU and GPU nodes optimized for both throughput and memory-intensive workloads. GPU nodes are configured for deep learning–based structure prediction and design tasks, while CPU nodes efficiently handle parallel reconstruction and refinement workflows.

Backed by a fault-tolerant 3.2 PB CephFS storage system with high-speed 100 GbE networking, EMCC enables efficient management of terabyte-scale Cryo-EM datasets and downstream analysis. All project data is regularly backed up to an off-site ZFS-based tier with snapshot recovery, ensuring data integrity and continuity.

Beyond Cryo-EM, EMCC supports protein crystallography, cryo-electron tomography, de novo protein design, molecular dynamics, and large language model applications—including retrieval-augmented generation (RAG) and scientific chatbots trained on in-house datasets.

Access to EMCC is available via secure SSH connections using AU’s jump-host infrastructure, the VDI.au.dk virtual desktop environment, or through AU's VPN service. Researchers can work interactively or submit batch jobs from remote terminals, workstations, or graphical desktop sessions.

In addition, EMCC hosts 13 high-end GPU-accelerated workstations in a dedicated on-site Graphics Room, allowing users to physically access powerful nodes for interactive 3D visualization, structural modeling, and real-time GPU workflows.

The cluster is available free of charge for pilot usage (up to 1 TB), with scalable project quotas (seats of 20 TB) and institutional support.

Research Fields & Associated Tools

Cryo-EM

cryoSPARCRELION

Tomography

IMODScipionAreTomo

X-ray Crystallography

XDSCCP4Phenix

Protein Design

AlphaFoldRFdiffusionBindCraftChai-1

Large Language Models

Open Source ChatbotsRAG chatbots

Your Area Here?

Fancy Pipeline 1Fancy pipeline 2

Quick Acess Guide

Login via VDI (no VPN needed)

  1. If your a registered user go to vdi.au.dk.
  2. Install the VMware Horizon Client when prompted, then launch it and add the server vdi.au.dk.
  3. Log in with your AU ID: au123456@uni.au.dk and your AU password.
    You’ll be asked for 2FA (same as other AU services).
  4. Pick any desktop that has “(public)” in its name.
  5. Inside the desktop, log in to EMCC with your EMCC username/password (not your AU password).

Direct SSH (alternative)

Terminal-only access using SSH or SFTP is also possible:

  1. If you’re off campus, connect to the AU VPN first.
  2. SSH or SFTP into one of the public EMCC access hosts (the same ones marked “(public)” in VDI).
    ssh/sftp emcc_username@<host-name>.bioxray.au.dk;
  3. You will be prompted for a password, which is your EMCC password, not your AU password.

Our Compute Nodes

CPU Nodes

Node CPU Type Cores Threads RAM Type RAM (GB) Network
vader1Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader2Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader3Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader4Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader5Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader6Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader7Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader8Xeon E5-2699 v33672DDR4-2133 MHz384100 GbE
vader9AMD EPYC 933464128DDR5-4800 MHz384100 GbE
vader10AMD EPYC 933464128DDR5-4800 MHz384100 GbE

GPU Nodes

Node CPU Type Cores Threads RAM Type RAM (GB) Network GPU Type
snoke4Xeon Gold 62263264DDR4-2933 MHz384100 GbE4× Tesla V100S
snoke5Intel Platinum 835864128DDR4-3200 MHz512100 GbE4× Tesla A100
snoke6Intel Platinum 835864128DDR4-3200 MHz1024100 GbE4× Tesla A100
snoke7AMD EPYC 933464128DDR5-4800 MHz768100 GbE4× Tesla A100
snoke8AMD EPYC 933464128DDR5-4800 MHz768100 GbE4× Tesla L40S
snoke9Intel Platinum 8562Y+64128DDR5-5600 MHz1024100 GbE4× Tesla L40S
snoke10Xeon Gold 6548N64128DDR5-5600 MHz1024100 GbE4× Tesla L40S

Infrastructure

Primary Storage

Name: EMCC Storage (CephFS)
Capacity: 3.2 PB raw (16 nodes, 292 disks)
Snapshots: Nightly (7-day retention)
Mounts: /u/<user>, /emcc/<project>
Access: CephFS, SSHFS
Data type: Non-sensitive research data (e.g., imaging, structural prediction)

Backup & Archiving

System: Remote ZFS pool (off-site, 1540-K13)
Backup: Nightly incremental (3-month retention)
Archive: Data optionally replicated to ERDA (30-day snapshots)
Data type: Non-sensitive research data (e.g., imaging, structural prediction)

Access & Security

Auth: IPA/IDM login (user/pass, SSH keys)
ACLs: POSIX permissions
Encryption: In transit only (Ceph msgr2, SSH/SFTP)
Compliance: GDPR & AU InfoSec Policy
Do not store sensitive or personal data.

Quotas & Costs

Default: 1 TB/user (free)
Project: 20 TB/project (paid)
Annual Cost: 50.000 DKK/year (after pilot)
Scalability: Online expansion (Ceph + ZFS)

Documentation

LabBook Wiki

Software‑stack documentation, step‑by‑step tutorials, general HPC info etc.

NB. AU credentials required.

Open Wiki

EMCC Chatbot

Chat with our open‑source language models (Gemma, Qwen, DeepSeek ∞) or ask the RAG assistants for help with Data Management Plans, CryoSPARC and more — all served from EMCC hardware.

NB. User account required!

Open Chatbot

EMCC Diskhogs

See the storage usage (Qutoa, % used, number of files etc) for each user/group account for both the EMCC cluster and the USERHOME. The stats are update each midnight

NB. The site is only accessible from inside AU network!

See The Diskhogs!

Main Contact

EMCC Support – Main contact

Click to email our help desk

Individual contacts

Rune Kidmose
Special Consultant

rtk@mbg.au.dk

Jesper Lykkegaard Karlsen
Chief Consultant

jelka@au.dk

Thomas Boesen
Cryo‑EM Manager

For EM-Facility related questions: thb@inano.au.dk