SlideShare a Scribd company logo
CHEP2012 – New York (USA), May 21-25 2012



   Design and implementation of a reliable and cost-effective
  cloud computing infrastructure: the INFN Napoli experience
        Vincenzo Capone, bRosario Esposito, bSilvio Pardi, b,cFrancesco Taurino, bGennaro Tortone
      a,b
                                                    Università degli Studi di Napoli Federico II – Napoli, Italy
                                                     a

                                            b
                                              INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy
                                             c
                                               CNR-SPIN - Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy
                              Email: ecapone@na.infn.it, resposit@na.infn.it, spardi@na.infn.it, taurino@na.infn.it, tortone@na.infn.it


Introduction
In this work, we describe an IaaS (Infrastructure as a Service) cloud
computing system, with high availability and redundancy features which
is currently in production at INFN-Napoli and ATLAS Tier-2 data centre.
The main goal we intended to achieve was a simplified method to
manage our computing resources and deliver reliable user services,
reusing existing hardware without incurring heavy costs. A combined
usage of virtualization and clustering technologies allowed us to
consolidate our services on a small number of physical machines,
reducing electric power costs. As a results of our efforts we developed a
complete solution for data and computing centers that can be easily
replicated using commodity hardware.


                                                                                                A snapshot of some the custom CLI tools.
                                                                 Hardware
                                                                 We started from commodity hardware we already owned, that was upgraded
                                                                 in order to fulfill the requested performance. In particular, three Dell
                                                                 PowerEdge 1950 rack servers have been used as VM executors, and two
                                                                 Dell PowerEdge 2950 as VM stores.
                                                                 All servers are equipped with dual Intel Xeon E5430, providing 8 cpu cores
                                                                 per server, with 8 Gbyte of RAM. The upgrades consisted in a 8 Gbyte RAM
                                                                 and 2 ports Ethernet NIC on hypervisors, 6 x 1.5 TByte SATA hard disk and a
                                                                 4 ports Ethernet NIC on both storage servers. The storage server disks are
                                                                 configured in RAID5 (dm-raid software mode), so the total available storage
                                                                 space is 7.5 Tbyte per server. Hypervisor servers have 2 x 500 Gbyte disks
                                                                 configured in RAID1. Furthermore a dedicated 24 gigabit ports Cisco Catalyst
                                                                 2960G switch was added to the hardware configuration to provide a
                                                                 dedicated storage LAN.

Network
The main requirements for the network serving our infrastructure are: performance, reliability and resiliency. To achieve these
goals, we set up a double path between every hypervisor and both storage servers, with two different switches involved, so that a
failure in one of them doesn’t impact on the execution of the Virtual Machines, whose disk images are hosted on the storage
servers. The Cisco Catalyst 6509 is the core switch of our science department network infrastructure, and every server is
connected to it via the onboard dual gigabit ethernet port, in LACP bonding mode, so to provide the necessary connectivity and the
sufficient bandwidth to the VMs: this link is in trunk mode, so that every VM can be connected to the desired VLAN. The second
switch (Cisco 2960G) is connected to the former via a 3 x 1 Gbit LACP bond link. A private VLAN hosts the data traffic between the
storage servers and the hypervisors; within this VLAN every storage server is connected with three gigabit links to the Cisco 2960G
and the fourth to the Cisco 6509, while every hypervisor is connected with one link to both switches; the multiple connection of the
servers to the two switches is achieved with the Adaptive Load Balancing mode. Within this topology, the Cisco 2960G is
completely dedicated to the network traffic of the storage VLAN, while the Cisco 6509 is used as access switch towards the LAN,
and as the redundant switch for the storage VLAN.


Software
 The OS used on all servers was initially Scientific Linux 5.5, with KVM as virtualization system. We selected KVM
 as the best architecture for virtualization on modern processors with fast hardware virtualization support (VT-x
 and NPT on Intel or AMD-V and EPT on AMD). After, we updated all servers to Scientific Linux 6.2 to use the
 new KVM version and KSM (Kernel Samepage Merging), a memory deduplication feature which enables more
 guests sharing to share the same memory pages of the host.


Storage
We chose GlusterFS as a fault-tolerant backend storage for virtual
machine images. GlusterFS is an open source, clustered file-
system for scaling the storage capacity of many servers to several
petabytes. It aggregates various storage servers or bricks over
Infiniband RDMA and/or TCP/IP interconnection into one large
parallel network file system. Key features of GlusterFS:
- Modular, stackable storage OS architecture
- Data stored in native formats
- No metadata – Elastic hashing
- Automatic file replication with self-healing.
In the GlusterFS world a volume is a logical collection of bricks, where each brick is
an export directory on a server in the trusted storage pool. Most of the Gluster
management operations happen on the volume. In our local setup we used a
replicated Gluster volume created on top of 2 servers to store virtual machine disk
images in qcow2 file format. Each storage server exports a storage brick which
consists of an ext3 file system built on a Linux software RAID5 array.                                               IOZone benchmark measuring disk r/w performance to GlusterFS:
                                                                                                                     1) storage server local array, 2) KVM host, 3) guest VM disk image
In the picture on the right are shown the disk performances under various use cases.


Features
We have developed some CLI scripts for day by day tasks on our private cloud in order to reduce administration efforts, like rapid
provisioning of guest, listing, rapid migration, load balancing and automatic migration and restart of VMs hosted on a failed
hypervisor.
With our deployment we’ve achieved all the goals we intended to: ease of management, high availability and fault tolerance. The
functional integrity of the whole cloud system is preserved even after the fault of multiple elements of the system: in fact, no other
effects, but the declining of the overall performance, happens after the failure of one of the two switches, one of the two storage
servers, all but one of the KVM hypervisors, even if all this happens at the same time.
In conclusion, our system has proved itself a solid and efficient solution, after more than one year of uninterrupted uptime, to deploy
all those services that don’t require a heavy load on the I/O subsystem, but that are a crucial element of a modern datacenter.


                                                               http://www.na.infn.it

More Related Content

PDF
PVOps Update
PDF
Building a Distributed Block Storage System on Xen
PDF
(Free and Net) BSD Xen Roadmap
PDF
Oracle rac 10g best practices
ODP
PPTX
Impact of Intel Optane Technology on HPC
PDF
Cascade lake-advanced-performance-press-deck
PDF
Xen PV Performance Status and Optimization Opportunities
PVOps Update
Building a Distributed Block Storage System on Xen
(Free and Net) BSD Xen Roadmap
Oracle rac 10g best practices
Impact of Intel Optane Technology on HPC
Cascade lake-advanced-performance-press-deck
Xen PV Performance Status and Optimization Opportunities

What's hot (20)

PPTX
Heterogeneous Computing on POWER - IBM and OpenPOWER technologies to accelera...
PPTX
Memory management in vx works
PDF
Xen in Linux 3.x (or PVOPS)
PDF
Building High Availability Clusters with SUSE Linux Enterprise High Availabil...
PDF
XS Boston 2008 XenLoop
PDF
KVM Tuning @ eBay
KEY
Finadmin virtualization2
PDF
Kvm performance optimization for ubuntu
PPTX
JetStor portfolio update final_2020-2021
ODP
Docker on Power Systems
PPTX
Openstorage with OpenStack, by Bradley
PDF
A guide of PostgreSQL on Kubernetes
PPTX
Realtime scheduling for virtual machines in SKT
PPTX
Revisiting CephFS MDS and mClock QoS Scheduler
PDF
Linux container & docker
PDF
A Survey of Performance Comparison between Virtual Machines and Containers
PDF
VMworld 2013: Cisco, VMware and Hyper-converged Solutions for the Enterprise....
PDF
Paravirtualized File Systems
PDF
General Bare-metal Provisioning Framework.pdf
PDF
1212312232
Heterogeneous Computing on POWER - IBM and OpenPOWER technologies to accelera...
Memory management in vx works
Xen in Linux 3.x (or PVOPS)
Building High Availability Clusters with SUSE Linux Enterprise High Availabil...
XS Boston 2008 XenLoop
KVM Tuning @ eBay
Finadmin virtualization2
Kvm performance optimization for ubuntu
JetStor portfolio update final_2020-2021
Docker on Power Systems
Openstorage with OpenStack, by Bradley
A guide of PostgreSQL on Kubernetes
Realtime scheduling for virtual machines in SKT
Revisiting CephFS MDS and mClock QoS Scheduler
Linux container & docker
A Survey of Performance Comparison between Virtual Machines and Containers
VMworld 2013: Cisco, VMware and Hyper-converged Solutions for the Enterprise....
Paravirtualized File Systems
General Bare-metal Provisioning Framework.pdf
1212312232
Ad

Viewers also liked (9)

PDF
AAI Nazionale
PDF
Open Security
PDF
AAI Locale
PPT
Unattended
PDF
APT per RPM
PDF
Proposte (informatiche) per il comune di Santa Maria Capua Vetere
PDF
Excelsior 2009
PDF
ClearOS - Linux Small Business Server
AAI Nazionale
Open Security
AAI Locale
Unattended
APT per RPM
Proposte (informatiche) per il comune di Santa Maria Capua Vetere
Excelsior 2009
ClearOS - Linux Small Business Server
Ad

Similar to Design and implementation of a reliable and cost-effective cloud computing infrastructure: the INFN Napoli experience (20)

PDF
How DreamHost builds a Public Cloud with OpenStack
PDF
How DreamHost builds a public cloud with OpenStack.pdf
PPTX
Open Source Cloud, Virtualization and Deployment Technologies
PPT
MyCloud for $100k
PDF
Practice and challenges from building IaaS
PDF
OpenStack Tutorial
PPT
How to Design a Scalable Private Cloud
PDF
XenSummit - 08/28/2012
PPTX
Cloud Computing Tools
PPTX
Cloud stack overview
PDF
Linuxtag 2012 - OpenNebula
ODP
Block Storage For VMs With Ceph
PDF
Building IaaS Clouds and the Art of Virtual Machine Management: A Practical G...
PDF
#VirtualDesignMaster 3 Challenge 3 – James Brown
PDF
Exploration of eucalyptus_v2.0
PDF
Red Hat Storage - Introduction to GlusterFS
PDF
Wicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
PPTX
Openstack Diablo Survey
How DreamHost builds a Public Cloud with OpenStack
How DreamHost builds a public cloud with OpenStack.pdf
Open Source Cloud, Virtualization and Deployment Technologies
MyCloud for $100k
Practice and challenges from building IaaS
OpenStack Tutorial
How to Design a Scalable Private Cloud
XenSummit - 08/28/2012
Cloud Computing Tools
Cloud stack overview
Linuxtag 2012 - OpenNebula
Block Storage For VMs With Ceph
Building IaaS Clouds and the Art of Virtual Machine Management: A Practical G...
#VirtualDesignMaster 3 Challenge 3 – James Brown
Exploration of eucalyptus_v2.0
Red Hat Storage - Introduction to GlusterFS
Wicked Easy Ceph Block Storage & OpenStack Deployment with Crowbar
Openstack Diablo Survey

More from Francesco Taurino (20)

ODP
La mia nuvola in azienda o a casa
PDF
Da 0 all'open per PA e PMI
PDF
LPPP - Linux Per le PMI (piccole e medie imprese) e la PA (Pubblica Amministr...
PDF
Francesco M. Taurino - Relazione tecnica e pubblicazioni
PDF
Presentazione del nuovo sito web del Comune di Santa Maria Capua Vetere
PDF
Open Source
PDF
Applicazioni open source
PDF
Una rete aziendale con Linux
PDF
Xen e OpenVirtuozzo
PDF
Back to Mainframe
PPT
PDF
Redhat RHCE Certification
PDF
PfSense Cluster
PDF
Monitoraggio ambientale a basso costo - 2
La mia nuvola in azienda o a casa
Da 0 all'open per PA e PMI
LPPP - Linux Per le PMI (piccole e medie imprese) e la PA (Pubblica Amministr...
Francesco M. Taurino - Relazione tecnica e pubblicazioni
Presentazione del nuovo sito web del Comune di Santa Maria Capua Vetere
Open Source
Applicazioni open source
Una rete aziendale con Linux
Xen e OpenVirtuozzo
Back to Mainframe
Redhat RHCE Certification
PfSense Cluster
Monitoraggio ambientale a basso costo - 2

Recently uploaded (20)

PPT
Teaching material agriculture food technology
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PPTX
MYSQL Presentation for SQL database connectivity
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
Encapsulation theory and applications.pdf
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
cuic standard and advanced reporting.pdf
PPTX
Cloud computing and distributed systems.
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PPTX
Big Data Technologies - Introduction.pptx
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
DOCX
The AUB Centre for AI in Media Proposal.docx
PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Chapter 3 Spatial Domain Image Processing.pdf
Teaching material agriculture food technology
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
MYSQL Presentation for SQL database connectivity
MIND Revenue Release Quarter 2 2025 Press Release
Understanding_Digital_Forensics_Presentation.pptx
Encapsulation theory and applications.pdf
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
cuic standard and advanced reporting.pdf
Cloud computing and distributed systems.
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
Big Data Technologies - Introduction.pptx
Agricultural_Statistics_at_a_Glance_2022_0.pdf
The AUB Centre for AI in Media Proposal.docx
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Advanced methodologies resolving dimensionality complications for autism neur...
Chapter 3 Spatial Domain Image Processing.pdf

Design and implementation of a reliable and cost-effective cloud computing infrastructure: the INFN Napoli experience

  • 1. CHEP2012 – New York (USA), May 21-25 2012 Design and implementation of a reliable and cost-effective cloud computing infrastructure: the INFN Napoli experience Vincenzo Capone, bRosario Esposito, bSilvio Pardi, b,cFrancesco Taurino, bGennaro Tortone a,b Università degli Studi di Napoli Federico II – Napoli, Italy a b INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy c CNR-SPIN - Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy Email: ecapone@na.infn.it, resposit@na.infn.it, spardi@na.infn.it, taurino@na.infn.it, tortone@na.infn.it Introduction In this work, we describe an IaaS (Infrastructure as a Service) cloud computing system, with high availability and redundancy features which is currently in production at INFN-Napoli and ATLAS Tier-2 data centre. The main goal we intended to achieve was a simplified method to manage our computing resources and deliver reliable user services, reusing existing hardware without incurring heavy costs. A combined usage of virtualization and clustering technologies allowed us to consolidate our services on a small number of physical machines, reducing electric power costs. As a results of our efforts we developed a complete solution for data and computing centers that can be easily replicated using commodity hardware. A snapshot of some the custom CLI tools. Hardware We started from commodity hardware we already owned, that was upgraded in order to fulfill the requested performance. In particular, three Dell PowerEdge 1950 rack servers have been used as VM executors, and two Dell PowerEdge 2950 as VM stores. All servers are equipped with dual Intel Xeon E5430, providing 8 cpu cores per server, with 8 Gbyte of RAM. The upgrades consisted in a 8 Gbyte RAM and 2 ports Ethernet NIC on hypervisors, 6 x 1.5 TByte SATA hard disk and a 4 ports Ethernet NIC on both storage servers. The storage server disks are configured in RAID5 (dm-raid software mode), so the total available storage space is 7.5 Tbyte per server. Hypervisor servers have 2 x 500 Gbyte disks configured in RAID1. Furthermore a dedicated 24 gigabit ports Cisco Catalyst 2960G switch was added to the hardware configuration to provide a dedicated storage LAN. Network The main requirements for the network serving our infrastructure are: performance, reliability and resiliency. To achieve these goals, we set up a double path between every hypervisor and both storage servers, with two different switches involved, so that a failure in one of them doesn’t impact on the execution of the Virtual Machines, whose disk images are hosted on the storage servers. The Cisco Catalyst 6509 is the core switch of our science department network infrastructure, and every server is connected to it via the onboard dual gigabit ethernet port, in LACP bonding mode, so to provide the necessary connectivity and the sufficient bandwidth to the VMs: this link is in trunk mode, so that every VM can be connected to the desired VLAN. The second switch (Cisco 2960G) is connected to the former via a 3 x 1 Gbit LACP bond link. A private VLAN hosts the data traffic between the storage servers and the hypervisors; within this VLAN every storage server is connected with three gigabit links to the Cisco 2960G and the fourth to the Cisco 6509, while every hypervisor is connected with one link to both switches; the multiple connection of the servers to the two switches is achieved with the Adaptive Load Balancing mode. Within this topology, the Cisco 2960G is completely dedicated to the network traffic of the storage VLAN, while the Cisco 6509 is used as access switch towards the LAN, and as the redundant switch for the storage VLAN. Software The OS used on all servers was initially Scientific Linux 5.5, with KVM as virtualization system. We selected KVM as the best architecture for virtualization on modern processors with fast hardware virtualization support (VT-x and NPT on Intel or AMD-V and EPT on AMD). After, we updated all servers to Scientific Linux 6.2 to use the new KVM version and KSM (Kernel Samepage Merging), a memory deduplication feature which enables more guests sharing to share the same memory pages of the host. Storage We chose GlusterFS as a fault-tolerant backend storage for virtual machine images. GlusterFS is an open source, clustered file- system for scaling the storage capacity of many servers to several petabytes. It aggregates various storage servers or bricks over Infiniband RDMA and/or TCP/IP interconnection into one large parallel network file system. Key features of GlusterFS: - Modular, stackable storage OS architecture - Data stored in native formats - No metadata – Elastic hashing - Automatic file replication with self-healing. In the GlusterFS world a volume is a logical collection of bricks, where each brick is an export directory on a server in the trusted storage pool. Most of the Gluster management operations happen on the volume. In our local setup we used a replicated Gluster volume created on top of 2 servers to store virtual machine disk images in qcow2 file format. Each storage server exports a storage brick which consists of an ext3 file system built on a Linux software RAID5 array. IOZone benchmark measuring disk r/w performance to GlusterFS: 1) storage server local array, 2) KVM host, 3) guest VM disk image In the picture on the right are shown the disk performances under various use cases. Features We have developed some CLI scripts for day by day tasks on our private cloud in order to reduce administration efforts, like rapid provisioning of guest, listing, rapid migration, load balancing and automatic migration and restart of VMs hosted on a failed hypervisor. With our deployment we’ve achieved all the goals we intended to: ease of management, high availability and fault tolerance. The functional integrity of the whole cloud system is preserved even after the fault of multiple elements of the system: in fact, no other effects, but the declining of the overall performance, happens after the failure of one of the two switches, one of the two storage servers, all but one of the KVM hypervisors, even if all this happens at the same time. In conclusion, our system has proved itself a solid and efficient solution, after more than one year of uninterrupted uptime, to deploy all those services that don’t require a heavy load on the I/O subsystem, but that are a crucial element of a modern datacenter. http://www.na.infn.it