SlideShare a Scribd company logo
The Why and How of HPC-Cloud Hybrids
with OpenStack
OpenStack Australia Day Melbourne
June, 2017
Lev Lafayette, HPC Support and Training Officer, University of Melbourne
lev.lafayette@unimelb.edu.au
1.0 Management Layer
1.1 HPC for performance
High-performance computing (HPC) is any computer system whose architecture allows for
above average performance. Main use case refers to compute clusters with a teamster
separation between head node and workers nodes and a high-speed interconnect acts as a
single system.
1.2 Clouds for flexibility.
Precursor with virtualised hardware. Cloud VMs always have lower performance than HPC.
Question is whether the flexibility is worth the overhead.
1.3 Hybrid HPC/Clouds.
University of Melbourne model, "the chimera". Cloud VMs deployed as HPC nodes and the
Freiburg University Model, "the cyborg", HPC nodes deploying Cloud VMs.
1.4 Reviewing user preferences and usage.
Users always want more of 'x'; real issue identified was queue times. Usage indicated a high
proportion of single-node jobs.
1.5 Review and Architecture.
Review discussed whether UoM needed HPC; architecture was to use existing NeCTAR
Research cloud with an expansion of general cloud compute provisioning and use of a smaller
"true HPC" system on bare metal nodes.
2.0 Physical Layer
2.1 Physical Partitions.
"Real" HPC is a mere c276 cores, 21 GB per core. 2 socket Intel E5-2643 v3 E5-2643,
3.4GHz CPU with 6-core per socket, 192GB memory, 2x 1.2TB SAS drives, 2x 40GbE
network. “Cloud” partitions is almost 400 virtual machines with over 3,000 2.3GHz Haswell
cores with 8GB per core and . There is also a GPU partition with Dual Nvidia Tesla K80s (big
expansion this year), and departmental partitions (water and ashley). Management and login
nodes are VMs as is I/O for transferring data.
2.2 Network.
System network includes: Cloud nodes Cisco Nexus 10Gbe TCP/IP 60 usec latency (mpi-
pingpong); Bare Metal Mellanox 2100 Cumulos Linux 40Gbe TCP/IP 6.85 usec latency and
then RDMA Ethernet 1.15 usec latency
2.3 Storage.
Mountpoints to home, projects (/project /home for user data & scripts, NetApp SAS aggregate
70TB usable) and applications directories across all nodes. Additional mountpoins to VicNode
Aspera Shares. Applications directory currently on management node, needs to be decoupled.
Bare metal nodes have /scratch shared storage for MPI jobs (Dell R730 with 14 x 800GB
mixed use SSDs providing 8TB of usable storage, NFS over RDMA)., /var/local/tmp for single
node jobs, pcie SSD 1.6TB.
3.0 Operating System and
Scheduler Layer
3.1 Red Hat Linux.
Scalable FOSS operating system, high performance, very well suited for research
applications. In November 2016 of the Top 500 Supercomputers worldwide, every single
machine used a "UNIX-like" operating system; and 99.6% used Linux.
3.2 Slurm Workload Manager.
Job schedulers and resource managers allow for unattended background tasks expressed as
batch jobs among the available resources; allows multicore, multinode, arrays, dependencies,
and interactive submissions. The scheduler provides for paramterisation of computer
resources, an automatic submission of execution tasks, and a notification system for incidents.
Slurm (originally Simple Linux Utility for Resource Management), developed by Lawrence
Livermore et al., is FOSS, and used by majority of world's top systems. Scalable, offers many
optional plugins, power-saving features, accounting features, etc. Divided into logical partitions
which correlate with hardware partitions.
3.3 Git, Gerrit, and Puppet.
Version control, paired systems administration, configuration management.
3.4 OpenStack Node Deployment.
Significant use of Nova (compute) service for provisioning and decommissioning of virtual
machines on demand.
4.0 Application Layer
4.1 Source Code and EasyBuild.
Source code provides better control over security updates, integration, development, and
much better performance. Absolutely essential for reproducibility in research environment.
EasyBuild makes source software installs easier with scripts containing specified compilation
blocks (e.g., configuremake, cmake etc) and specified toolchains (GCC, Intel etc) and
environment modules (LMod). Modulefiles allow for dynamic changes to a user's environment
and ease with multiple versions of software applications on a system.
4.2 Compilers, Scripting Languages, and Applications.
Usual range of suspects; Intel and GCC, for compilers (and a little bit of PGI), Python Ruby,
and Perl for scripting languages, OpenMPI wrappers. Major applications include: MATLAB,
Gaussian, NAMD, R, OpenFOAM, Octave etc.
Almost 1,000 applications/versions installed from source, plus packages.
4.3 Containers with Singularity.
A container in a cloud virtual machine on an HPC! Wait, what?
5.0 User Layer
5.1 Karaage.
Spartan uses its own LDAP authentication that is tied to the university Security Assertion
Markup Language (SAML). Users on Spartan must belong to a project. Projects must be led
by a University of Melbourne researcher (the "Principal Investigator") and are subject to
approval by the Head of Research Compute Services. Participants in a project can be
researchers or research support staff from anywhere. Karaage is Django-based application for
user, project, and cluster reporting and management.
5.2 Freshdesk.
OMG Users!
5.3 Online Instructions and Training.
Many users (even post-doctoral researchers) require basic training in Linux command line, a
requisite skill for HPC use. Extensive training programme for researchers available using
andragogical methods, including day-long courses in “Introduction to Linux and HPC Using
Spartan”, “Linux Shell Scripting for High Performance Computing”, and “Parallel Programming
On Spartan”.
Documentation online (Github, Website, and man pages) and plenty of Slurm examples on
system.
6.0 Future Developments
6.1 Cloudbursting with Azure.
Slurm allows cloudbursting via the powersave feature; successfully experiments (and bug
discovery) within the NeCTAR research cloud.
About to add Azure through same login node. Does not mount applications directory; wrap
necessary data for transfer in script.
6.2 GPU Expansion.
Plans for a significant increase in the GPU allocation.
6.3 Test cluster (Thespian).
Everyone has a test environment, some people also have a production and a test
environment.
Test nodes already exist for Cloud and Physical partitions. Replicate management and login
nodes.
6.3 New Architectures
New architectures can be added to the system with separate build node (another VM) and with
software built for that architecture. Don't need an entirely new system.
The Why and How of HPC-Cloud Hybrids with OpenStack - Lev Lafayette, University of Melbourne

More Related Content

PPT
OpenStack - An Overview
PPTX
WAN & LAN Cluster with Diagrams and OSI explanation
PDF
HPC Best Practices: Application Performance Optimization
PDF
Microkernel design
DOC
PDF
IBM Data Centric Systems & OpenPOWER
PPT
Cluster Computers
PPTX
Optimising nfv service chains on open stack using docker
OpenStack - An Overview
WAN & LAN Cluster with Diagrams and OSI explanation
HPC Best Practices: Application Performance Optimization
Microkernel design
IBM Data Centric Systems & OpenPOWER
Cluster Computers
Optimising nfv service chains on open stack using docker

What's hot (20)

PDF
Computer_Clustering_Technologies
PDF
The State of Linux Containers
PDF
Dell Lustre Storage Architecture Presentation - MBUG 2016
PPTX
April 2016 HUG: The latest of Apache Hadoop YARN and running your docker apps...
PDF
Application layer
PPT
Cluster Computing Seminar.
ODP
Systems Support for Many Task Computing
PPTX
EAS Data Flow lessons learnt
PDF
Teaching Apache Spark: Demonstrations on the Databricks Cloud Platform
PPTX
Beowulf cluster
PDF
Blazing Fast Lustre Storage
PPTX
Cluster computer
PDF
Enea Enabling Real-Time in Linux Whitepaper
PPT
Cluster Computing
PDF
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
PDF
CloudLightning and the OPM-based Use Case
PPT
tittle
PPTX
DOE Magellan OpenStack user story
PDF
Oracle rac 10g best practices
PDF
Docker Application to Scientific Computing
Computer_Clustering_Technologies
The State of Linux Containers
Dell Lustre Storage Architecture Presentation - MBUG 2016
April 2016 HUG: The latest of Apache Hadoop YARN and running your docker apps...
Application layer
Cluster Computing Seminar.
Systems Support for Many Task Computing
EAS Data Flow lessons learnt
Teaching Apache Spark: Demonstrations on the Databricks Cloud Platform
Beowulf cluster
Blazing Fast Lustre Storage
Cluster computer
Enea Enabling Real-Time in Linux Whitepaper
Cluster Computing
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
CloudLightning and the OPM-based Use Case
tittle
DOE Magellan OpenStack user story
Oracle rac 10g best practices
Docker Application to Scientific Computing
Ad

Similar to The Why and How of HPC-Cloud Hybrids with OpenStack - Lev Lafayette, University of Melbourne (20)

PDF
Japan's post K Computer
PPTX
HPC and cloud distributed computing, as a journey
PPT
Pacemaker+DRBD
PDF
Openstack_administration
PDF
Introduction to Apache Mesos and DC/OS
PPTX
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
PPTX
Microx - A Unix like kernel for Embedded Systems written from scratch.
PDF
Accelerate Big Data Processing with High-Performance Computing Technologies
PPT
Parallel_and_Cluster_Computing.ppt
PPTX
OS for AI: Elastic Microservices & the Next Gen of ML
PPTX
Clustering by AKASHMSHAH
PPTX
djypllh5r1gjbaekxgwv-signature-cc6692615bbc55079760b9b0c6636bc58ec509cd0446cb...
PDF
Rha cluster suite wppdf
PPT
Zou Layered VO PDCAT2008 V0.5 Concise
PPTX
M0339_v1_6977127809 (1).pptx
PPTX
Accelerating TensorFlow with RDMA for high-performance deep learning
PDF
From Rack scale computers to Warehouse scale computers
PDF
Red Hat multi-cluster management & what's new in OpenShift
PPTX
Cloud infrastructure, Virtualization tec
PDF
Beyond static configuration
Japan's post K Computer
HPC and cloud distributed computing, as a journey
Pacemaker+DRBD
Openstack_administration
Introduction to Apache Mesos and DC/OS
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
Microx - A Unix like kernel for Embedded Systems written from scratch.
Accelerate Big Data Processing with High-Performance Computing Technologies
Parallel_and_Cluster_Computing.ppt
OS for AI: Elastic Microservices & the Next Gen of ML
Clustering by AKASHMSHAH
djypllh5r1gjbaekxgwv-signature-cc6692615bbc55079760b9b0c6636bc58ec509cd0446cb...
Rha cluster suite wppdf
Zou Layered VO PDCAT2008 V0.5 Concise
M0339_v1_6977127809 (1).pptx
Accelerating TensorFlow with RDMA for high-performance deep learning
From Rack scale computers to Warehouse scale computers
Red Hat multi-cluster management & what's new in OpenShift
Cloud infrastructure, Virtualization tec
Beyond static configuration
Ad

More from OpenStack (20)

PDF
Swinburne University of Technology - Shunde Zhang & Kieran Spear, Aptira
PDF
Related OSS Projects - Peter Rowe, Flexera Software
PDF
Supercomputing by API: Connecting Modern Web Apps to HPC
PDF
Federation and Interoperability in the Nectar Research Cloud
PDF
Simplifying the Move to OpenStack
PDF
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
PDF
Migrating your infrastructure to OpenStack - Avi Miller, Oracle
PDF
A glimpse into an industry Cloud using Open Source Technologies - Adrian Koh,...
PDF
Enabling OpenStack for Enterprise - Tarso Dos Santos, Veritas
PDF
Understanding blue store, Ceph's new storage backend - Tim Serong, SUSE
PDF
OpenStack Networks the Web-Scale Way - Scott Laffer, Cumulus Networks
PPTX
Diving in the desert: A quick overview into OpenStack Sahara capabilities - A...
PDF
Building a GPU-enabled OpenStack Cloud for HPC - Blair Bethwaite, Monash Univ...
PDF
OpenStack and Red Hat: How we learned to adapt with our customers in a maturi...
PDF
Meshing OpenStack and Bare Metal Networks with EVPN - David Iles, Mellanox Te...
PPTX
Ironically, Infrastructure Doesn't Matter - Quinton Anderson, Commonwealth Ba...
PDF
Traditional Enterprise to OpenStack Cloud - An Unexpected Journey
PDF
Building a GPU-enabled OpenStack Cloud for HPC - Lance Wilson, Monash University
PDF
Monitoring Uptime on the NeCTAR Research Cloud - Andy Botting, University of ...
PPTX
Containers and OpenStack: Marc Van Hoof, Kumulus: Containers and OpenStack
Swinburne University of Technology - Shunde Zhang & Kieran Spear, Aptira
Related OSS Projects - Peter Rowe, Flexera Software
Supercomputing by API: Connecting Modern Web Apps to HPC
Federation and Interoperability in the Nectar Research Cloud
Simplifying the Move to OpenStack
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
Migrating your infrastructure to OpenStack - Avi Miller, Oracle
A glimpse into an industry Cloud using Open Source Technologies - Adrian Koh,...
Enabling OpenStack for Enterprise - Tarso Dos Santos, Veritas
Understanding blue store, Ceph's new storage backend - Tim Serong, SUSE
OpenStack Networks the Web-Scale Way - Scott Laffer, Cumulus Networks
Diving in the desert: A quick overview into OpenStack Sahara capabilities - A...
Building a GPU-enabled OpenStack Cloud for HPC - Blair Bethwaite, Monash Univ...
OpenStack and Red Hat: How we learned to adapt with our customers in a maturi...
Meshing OpenStack and Bare Metal Networks with EVPN - David Iles, Mellanox Te...
Ironically, Infrastructure Doesn't Matter - Quinton Anderson, Commonwealth Ba...
Traditional Enterprise to OpenStack Cloud - An Unexpected Journey
Building a GPU-enabled OpenStack Cloud for HPC - Lance Wilson, Monash University
Monitoring Uptime on the NeCTAR Research Cloud - Andy Botting, University of ...
Containers and OpenStack: Marc Van Hoof, Kumulus: Containers and OpenStack

Recently uploaded (20)

PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PPTX
Cloud computing and distributed systems.
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
KodekX | Application Modernization Development
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
CIFDAQ's Market Insight: SEC Turns Pro Crypto
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Reach Out and Touch Someone: Haptics and Empathic Computing
Cloud computing and distributed systems.
Building Integrated photovoltaic BIPV_UPV.pdf
Advanced methodologies resolving dimensionality complications for autism neur...
Encapsulation_ Review paper, used for researhc scholars
KodekX | Application Modernization Development
Chapter 3 Spatial Domain Image Processing.pdf
20250228 LYD VKU AI Blended-Learning.pptx
Understanding_Digital_Forensics_Presentation.pptx
CIFDAQ's Market Insight: SEC Turns Pro Crypto
Mobile App Security Testing_ A Comprehensive Guide.pdf
NewMind AI Weekly Chronicles - August'25 Week I
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Per capita expenditure prediction using model stacking based on satellite ima...
Network Security Unit 5.pdf for BCA BBA.
Dropbox Q2 2025 Financial Results & Investor Presentation
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Diabetes mellitus diagnosis method based random forest with bat algorithm
“AI and Expert System Decision Support & Business Intelligence Systems”
How UI/UX Design Impacts User Retention in Mobile Apps.pdf

The Why and How of HPC-Cloud Hybrids with OpenStack - Lev Lafayette, University of Melbourne

  • 1. The Why and How of HPC-Cloud Hybrids with OpenStack OpenStack Australia Day Melbourne June, 2017 Lev Lafayette, HPC Support and Training Officer, University of Melbourne lev.lafayette@unimelb.edu.au
  • 2. 1.0 Management Layer 1.1 HPC for performance High-performance computing (HPC) is any computer system whose architecture allows for above average performance. Main use case refers to compute clusters with a teamster separation between head node and workers nodes and a high-speed interconnect acts as a single system. 1.2 Clouds for flexibility. Precursor with virtualised hardware. Cloud VMs always have lower performance than HPC. Question is whether the flexibility is worth the overhead. 1.3 Hybrid HPC/Clouds. University of Melbourne model, "the chimera". Cloud VMs deployed as HPC nodes and the Freiburg University Model, "the cyborg", HPC nodes deploying Cloud VMs. 1.4 Reviewing user preferences and usage. Users always want more of 'x'; real issue identified was queue times. Usage indicated a high proportion of single-node jobs. 1.5 Review and Architecture. Review discussed whether UoM needed HPC; architecture was to use existing NeCTAR Research cloud with an expansion of general cloud compute provisioning and use of a smaller "true HPC" system on bare metal nodes.
  • 3. 2.0 Physical Layer 2.1 Physical Partitions. "Real" HPC is a mere c276 cores, 21 GB per core. 2 socket Intel E5-2643 v3 E5-2643, 3.4GHz CPU with 6-core per socket, 192GB memory, 2x 1.2TB SAS drives, 2x 40GbE network. “Cloud” partitions is almost 400 virtual machines with over 3,000 2.3GHz Haswell cores with 8GB per core and . There is also a GPU partition with Dual Nvidia Tesla K80s (big expansion this year), and departmental partitions (water and ashley). Management and login nodes are VMs as is I/O for transferring data. 2.2 Network. System network includes: Cloud nodes Cisco Nexus 10Gbe TCP/IP 60 usec latency (mpi- pingpong); Bare Metal Mellanox 2100 Cumulos Linux 40Gbe TCP/IP 6.85 usec latency and then RDMA Ethernet 1.15 usec latency 2.3 Storage. Mountpoints to home, projects (/project /home for user data & scripts, NetApp SAS aggregate 70TB usable) and applications directories across all nodes. Additional mountpoins to VicNode Aspera Shares. Applications directory currently on management node, needs to be decoupled. Bare metal nodes have /scratch shared storage for MPI jobs (Dell R730 with 14 x 800GB mixed use SSDs providing 8TB of usable storage, NFS over RDMA)., /var/local/tmp for single node jobs, pcie SSD 1.6TB.
  • 4. 3.0 Operating System and Scheduler Layer 3.1 Red Hat Linux. Scalable FOSS operating system, high performance, very well suited for research applications. In November 2016 of the Top 500 Supercomputers worldwide, every single machine used a "UNIX-like" operating system; and 99.6% used Linux. 3.2 Slurm Workload Manager. Job schedulers and resource managers allow for unattended background tasks expressed as batch jobs among the available resources; allows multicore, multinode, arrays, dependencies, and interactive submissions. The scheduler provides for paramterisation of computer resources, an automatic submission of execution tasks, and a notification system for incidents. Slurm (originally Simple Linux Utility for Resource Management), developed by Lawrence Livermore et al., is FOSS, and used by majority of world's top systems. Scalable, offers many optional plugins, power-saving features, accounting features, etc. Divided into logical partitions which correlate with hardware partitions. 3.3 Git, Gerrit, and Puppet. Version control, paired systems administration, configuration management. 3.4 OpenStack Node Deployment. Significant use of Nova (compute) service for provisioning and decommissioning of virtual machines on demand.
  • 5. 4.0 Application Layer 4.1 Source Code and EasyBuild. Source code provides better control over security updates, integration, development, and much better performance. Absolutely essential for reproducibility in research environment. EasyBuild makes source software installs easier with scripts containing specified compilation blocks (e.g., configuremake, cmake etc) and specified toolchains (GCC, Intel etc) and environment modules (LMod). Modulefiles allow for dynamic changes to a user's environment and ease with multiple versions of software applications on a system. 4.2 Compilers, Scripting Languages, and Applications. Usual range of suspects; Intel and GCC, for compilers (and a little bit of PGI), Python Ruby, and Perl for scripting languages, OpenMPI wrappers. Major applications include: MATLAB, Gaussian, NAMD, R, OpenFOAM, Octave etc. Almost 1,000 applications/versions installed from source, plus packages. 4.3 Containers with Singularity. A container in a cloud virtual machine on an HPC! Wait, what?
  • 6. 5.0 User Layer 5.1 Karaage. Spartan uses its own LDAP authentication that is tied to the university Security Assertion Markup Language (SAML). Users on Spartan must belong to a project. Projects must be led by a University of Melbourne researcher (the "Principal Investigator") and are subject to approval by the Head of Research Compute Services. Participants in a project can be researchers or research support staff from anywhere. Karaage is Django-based application for user, project, and cluster reporting and management. 5.2 Freshdesk. OMG Users! 5.3 Online Instructions and Training. Many users (even post-doctoral researchers) require basic training in Linux command line, a requisite skill for HPC use. Extensive training programme for researchers available using andragogical methods, including day-long courses in “Introduction to Linux and HPC Using Spartan”, “Linux Shell Scripting for High Performance Computing”, and “Parallel Programming On Spartan”. Documentation online (Github, Website, and man pages) and plenty of Slurm examples on system.
  • 7. 6.0 Future Developments 6.1 Cloudbursting with Azure. Slurm allows cloudbursting via the powersave feature; successfully experiments (and bug discovery) within the NeCTAR research cloud. About to add Azure through same login node. Does not mount applications directory; wrap necessary data for transfer in script. 6.2 GPU Expansion. Plans for a significant increase in the GPU allocation. 6.3 Test cluster (Thespian). Everyone has a test environment, some people also have a production and a test environment. Test nodes already exist for Cloud and Physical partitions. Replicate management and login nodes. 6.3 New Architectures New architectures can be added to the system with separate build node (another VM) and with software built for that architecture. Don't need an entirely new system.