SlideShare a Scribd company logo
Roberto Franchini 
franchini@celi.it 
Codemotion Milano 
29/11/2014
GlusterFS 
A scalable distributed 
file system
whoami(1) 
15 years of experience, proud to be a programmer 
Writes software for information extraction, nlp, opinion mining 
(@scale ), and a lot of other buzzwords 
Implements scalable architectures 
Plays with servers 
Member of the JUG-Torino coordination team 
franchini@celi.it 
http://www.celi.it http://www.blogmeter.it 
github.com/robfrank github.com/uim-celi 
twitter.com/robfrankie linkedin.com/in/robfrank
The problem 
Identify a distributed and scalable 
file system 
for today's and tomorrow's 
Big Data
Once upon a time 
2008: One nfs share 
1,5TB ought to be enough for anybody 
2010: Herd of shares 
(1,5TB x N) ought to be enough for anybody 
Nobody couldn’t stop the data flood 
It was the time for something new
Requirements 
Can be enlarged on demand 
No dedicated HW 
OS is preferred and trusted 
No specialized API 
No specialized Kernel 
POSIX compliance 
Zilions of big and small files 
No NAS or SAN (€€€€€)
Clustered Scale-out General Purpose Storage 
Platform 
− POSIX-y Distributed File System 
− ...and so much more 
Built on commodity systems 
− x86_64 Linux ++ 
− POSIX filesystems underneath (XFS, 
EXT4) 
No central metadata Server (NO SPOF) 
Modular architecture for scale and functionality
Common 
use cases 
Large Scale File Server 
Media / Content Distribution Network (CDN) 
Backup / Archive / Disaster Recovery (DR) 
High Performance Computing (HPC) 
Infrastructure as a Service (IaaS) storage layer 
Database offload (blobs) 
Unified Object Store + File Access
Features 
ACL and Quota support 
Fault-tolerance 
Peer to peer 
Self-healing 
Fast setup up 
Enlarge on demand 
Shrink on demand 
Snapshot 
On premise phisical or virtual 
On cloud
Architecture
Architecture 
Peer / Node 
− cluster servers (glusterfs server) 
− Runs the gluster daemons and participates in volumes 
Brick 
− A filesystem mountpoint on servers 
− A unit of storage used as a capacity building block
Bricks on a node
Architecture 
Translator 
− Logic between bricks or subvolume that generate a 
subvolume with certain characteristic 
− distribute, replica, stripe are special translators to 
generate simil-RAID configuration 
− perfomance translators 
Volume 
− Bricks combined and passed through translators 
− Ultimately, what's presented to the end user
Volume
Volume types
Distributed 
The default configuration 
Files “evenly” spread across bricks 
Similar to file-level RAID 0 
Server/Disk failure could be catastrophic
Distributed
Replicated 
Files written synchronously to replica peers 
Files read synchronously, 
but ultimately serviced by the first responder 
Similar to file-level RAID 1
Replicated
Distributed + replicated 
Distribued + replicated 
Similar to file-level RAID 10 
Most used layout
Distributed replicated
Striped 
Individual files split among bricks (sparse files) 
Similar to block-level RAID 0 
Limited Use Cases 
HPC Pre/Post Processing 
File size exceeds brick size
Striped
Moving parts
Components 
glusterd 
Management daemon 
One instance on each GlusterFS server 
Interfaced through gluster CLI 
glusterfsd 
GlusterFS brick daemon 
One process for each brick on each server 
Managed by glusterd
Components 
glusterfs 
Volume service daemon 
One process for each volume service 
NFS server, FUSE client, Self-Heal, Quota, ... 
mount.glusterfs 
FUSE native client mount extension 
gluster 
Gluster Console Manager (CLI)
Clients
Clients: native 
FUSE kernel module allows the filesystem to be built and 
operated entirely in userspace 
Specify mount to any GlusterFS server 
Native Client fetches volfile from mount server, then 
communicates directly with all nodes to access data 
Recommended for high concurrency and high write 
performance 
Load is inherently balanced across distributed volumes
Clients:NFS 
Standard NFS v3 clients 
Standard automounter is supported 
Mount to any server, or use a load balancer 
GlusterFS NFS server includes Network Lock Manager 
(NLM) to synchronize locks across clients 
Better performance for reading many small files from a 
single client 
Load balancing must be managed externally
Clients: libgfapi 
Introduced with GlusterFS 3.4 
User-space library for accessing data in GlusterFS 
Filesystem-like API 
Runs in application process 
no FUSE, no copies, no context switches 
...but same volfiles, translators, etc.
Clients: SMB/CIFS 
In GlusterFS 3.4 – Samba + libgfapi 
No need for local native client mount & re-export 
Significant performance improvements with FUSE 
removed from the equation 
Must be setup on each server you wish to connect to via 
CIFS 
CTDB is required for Samba clustering
Clients: HDFS 
Access data within and outside of Hadoop 
No HDFS name node single point of failure / bottleneck 
Seamless replacement for HDFS 
Scales with the massive growth of big data
Scalability
Under the hood 
Elastic Hash Algorithm 
No central metadata 
No Performance Bottleneck 
Eliminates risk scenarios 
Location hashed intelligently on filename 
Unique identifiers (GFID), similar to md5sum
Scalability 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
Gluster Server 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
Gluster Server 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
3TB 
Gluster Server 
Scale out performance and availability 
Scale out capacitry
Scalability 
Add disks to servers to increase storage size 
Add servers to increase bandwidth and storage size 
Add servers to increase availability (replica factor)
What we do with glusterFS
What we do with GFS 
Daily production of more than 10GB of Lucene inverted 
indexes stored on glusterFS 
more than 200GB/month 
Search stored indexes to extract different sets of 
documents for every customers 
YES: we open indexes directly on storage 
(it's POSIX!!!)
2010: first installation 
Version 3.0.x 
8 (not dedicated) servers 
Distributed replicated 
No bound on brick size (!!!!) 
Ca 4TB avaliable 
NOTE: stuck to 3.0.x until 2012 due to problems on 3.1 and 
3.2 series, then RH acquired gluster (RH Storage)
2012: (little) cluster 
New installation, version 3.3.2 
4TB available on 8 servers (DELL c5000) 
still not dedicated 
1 brick per server limited to 1TB 
2TB-raid 1 on each server 
Still in production
2012: enlarge 
New installation, upgrade to 3.3.x 
6TB available on 12 servers (still not dedicated) 
Enlarged to 9TB on 18 servers 
Bricks size bounded AND unbounded
2013: fail 
18 not dedicated servers: too much 
18 bricks of different sizes 
2 big down due to bricks out of space 
Didn’t restart after a move 
but… 
All data were recovered 
(files are scattered on bricks, read from them!)
2014: consolidate 
2 dedicated servers 
12 x 3TB SAS raid6 
4 bricks per server 
28 TB available 
distributed replicated 
4x1Gb bonded NIC 
ca 40 clients (FUSE) (other 
servers)
Consolidate 
brick 1 
brick 2 
brick 3 
brick 4 
Gluster Server 1 
brick 1 
brick 2 
brick 3 
brick 4 
Gluster Server 2
Scale up 
brick 11 
brick 12 
brick 13 
brick 31 
Gluster Server 1 
brick 21 
brick 22 
brick 32 
brick 24 
Gluster Server 2 
brick 31 
brick 32 
brick 23 
brick 14 
Gluster Server 3
Do 
Dedicated server (phisical or virtual) 
RAID 6 or RAID 10 (with small files) 
Multiple bricks of same size 
Plan to scale
Do not 
Multi purpose server 
Bricks of different size 
Very small files 
Write to bricks
Some raw tests 
read 
Total transferred file size: 23.10G bytes 
43.46M bytes/sec 
write 
Total transferred file size: 23.10G bytes 
38.53M bytes/sec
Raw tests 
NOTE: ran in production under heavy load, no 
clean test room
Resources 
http://www.gluster.org/ 
https://access.redhat.com/documentation/en- 
US/Red_Hat_Storage/ 
https://github.com/gluster 
http://www.redhat.com/products/storage-server/ 
http://joejulian.name/blog/category/glusterfs/ 
http://jread.us/2013/06/one-petabyte-red-hat-storage-and-glusterfs- 
project-overview/
Thank you!
GlusterFs: a scalable file system for today's and tomorrow's big data
Roberto Franchini 
franchini@celi.it

More Related Content

PDF
MariaDB Performance Tuning and Optimization
PDF
Optimizing MariaDB for maximum performance
PDF
PostgreSQL Deep Internal
PDF
Ceph Object Storage Reference Architecture Performance and Sizing Guide
PDF
NY Meetup: Scaling MariaDB with Maxscale
PDF
Mastering PostgreSQL Administration
 
PDF
Présentation de Apache Zookeeper
PDF
Solving PostgreSQL wicked problems
MariaDB Performance Tuning and Optimization
Optimizing MariaDB for maximum performance
PostgreSQL Deep Internal
Ceph Object Storage Reference Architecture Performance and Sizing Guide
NY Meetup: Scaling MariaDB with Maxscale
Mastering PostgreSQL Administration
 
Présentation de Apache Zookeeper
Solving PostgreSQL wicked problems

What's hot (20)

PPTX
HBase Accelerated: In-Memory Flush and Compaction
PPTX
APACHE KAFKA / Kafka Connect / Kafka Streams
PPTX
Data models in NoSQL
PPTX
Kubernetes best practices with GKE
PPTX
Hive, Presto, and Spark on TPC-DS benchmark
PDF
Ansible - Hands on Training
PPTX
Ceph Intro and Architectural Overview by Ross Turk
PPTX
High Performance, High Reliability Data Loading on ClickHouse
PDF
MySQL InnoDB Cluster - New Features in 8.0 Releases - Best Practices
PDF
MySQL Database Architectures - MySQL InnoDB ClusterSet 2021-11
PDF
AmebaのMongoDB活用事例
PDF
Oracle RAC - Roadmap for New Features
PDF
MySQL High Availability with Group Replication
PPTX
HDFSネームノードのHAについて #hcj13w
PDF
Galera Cluster for MySQL vs MySQL (NDB) Cluster: A High Level Comparison
PDF
Data Engineer's Lunch #83: Strategies for Migration to Apache Iceberg
PDF
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
PDF
MySQL Performance Tuning. Part 1: MySQL Configuration (includes MySQL 5.7)
PDF
The Full MySQL and MariaDB Parallel Replication Tutorial
PDF
Learning postgresql
HBase Accelerated: In-Memory Flush and Compaction
APACHE KAFKA / Kafka Connect / Kafka Streams
Data models in NoSQL
Kubernetes best practices with GKE
Hive, Presto, and Spark on TPC-DS benchmark
Ansible - Hands on Training
Ceph Intro and Architectural Overview by Ross Turk
High Performance, High Reliability Data Loading on ClickHouse
MySQL InnoDB Cluster - New Features in 8.0 Releases - Best Practices
MySQL Database Architectures - MySQL InnoDB ClusterSet 2021-11
AmebaのMongoDB活用事例
Oracle RAC - Roadmap for New Features
MySQL High Availability with Group Replication
HDFSネームノードのHAについて #hcj13w
Galera Cluster for MySQL vs MySQL (NDB) Cluster: A High Level Comparison
Data Engineer's Lunch #83: Strategies for Migration to Apache Iceberg
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
MySQL Performance Tuning. Part 1: MySQL Configuration (includes MySQL 5.7)
The Full MySQL and MariaDB Parallel Replication Tutorial
Learning postgresql
Ad

Viewers also liked (8)

PDF
Codemotion Rome 2015. GlusterFS
PDF
What the hell is your software doing at runtime?
PPTX
Where are yours vertexes and what are they talking about?
PDF
Java application monitoring with Dropwizard Metrics and graphite
PDF
Java 8: le nuove-interfacce di Ezio Sperduto
PDF
Redis for duplicate detection on real time stream
PDF
Metodi asincroni in spring
PPTX
DevExperience - The Dark Side of Microservices
Codemotion Rome 2015. GlusterFS
What the hell is your software doing at runtime?
Where are yours vertexes and what are they talking about?
Java application monitoring with Dropwizard Metrics and graphite
Java 8: le nuove-interfacce di Ezio Sperduto
Redis for duplicate detection on real time stream
Metodi asincroni in spring
DevExperience - The Dark Side of Microservices
Ad

Similar to GlusterFs: a scalable file system for today's and tomorrow's big data (20)

PDF
Gluster FS a filesistem for Big Data | Roberto Franchini - Codemotion Rome 2015
PDF
Gluster intro-tdose
ODP
Gluster intro-tdose
PDF
Introducing gluster filesystem by aditya
ODP
GlusterFS Architecture - June 30, 2011 Meetup
PPTX
Celi @Codemotion 2014 - Roberto Franchini GlusterFS
ODP
The Future of GlusterFS and Gluster.org
ODP
Gluster fs architecture_future_directions_tlv
PDF
Gluster fs architecture_&_roadmap-vijay_bellur-linuxcon_eu_2013
PDF
Gluster fs architecture_future_directions_tlv
ODP
Glusterfs for sysadmins-justin_clift
PDF
vBACD - Distributed Petabyte-Scale Cloud Storage with GlusterFS - 2/28
PDF
GlusterFS Update and OpenStack Integration
PDF
The Future of GlusterFS and Gluster.org
PDF
State of the_gluster_-_lceu
PDF
Glusterfs and openstack
PDF
GlusterFS Presentation FOSSCOMM2013 HUA, Athens, GR
ODP
GlusterFs Architecture & Roadmap - LinuxCon EU 2013
PDF
Gluster fs tutorial part 2 gluster and big data- gluster for devs and sys ...
PDF
GlusterFS Talk for CentOS Dojo Bangalore
Gluster FS a filesistem for Big Data | Roberto Franchini - Codemotion Rome 2015
Gluster intro-tdose
Gluster intro-tdose
Introducing gluster filesystem by aditya
GlusterFS Architecture - June 30, 2011 Meetup
Celi @Codemotion 2014 - Roberto Franchini GlusterFS
The Future of GlusterFS and Gluster.org
Gluster fs architecture_future_directions_tlv
Gluster fs architecture_&_roadmap-vijay_bellur-linuxcon_eu_2013
Gluster fs architecture_future_directions_tlv
Glusterfs for sysadmins-justin_clift
vBACD - Distributed Petabyte-Scale Cloud Storage with GlusterFS - 2/28
GlusterFS Update and OpenStack Integration
The Future of GlusterFS and Gluster.org
State of the_gluster_-_lceu
Glusterfs and openstack
GlusterFS Presentation FOSSCOMM2013 HUA, Athens, GR
GlusterFs Architecture & Roadmap - LinuxCon EU 2013
Gluster fs tutorial part 2 gluster and big data- gluster for devs and sys ...
GlusterFS Talk for CentOS Dojo Bangalore

Recently uploaded (20)

PDF
Addressing The Cult of Project Management Tools-Why Disconnected Work is Hold...
PDF
Adobe Illustrator 28.6 Crack My Vision of Vector Design
PPTX
CHAPTER 2 - PM Management and IT Context
PDF
Design an Analysis of Algorithms II-SECS-1021-03
PDF
Nekopoi APK 2025 free lastest update
PDF
top salesforce developer skills in 2025.pdf
PDF
Designing Intelligence for the Shop Floor.pdf
PPTX
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
PDF
Cost to Outsource Software Development in 2025
PDF
Which alternative to Crystal Reports is best for small or large businesses.pdf
PDF
iTop VPN Free 5.6.0.5262 Crack latest version 2025
PDF
Why TechBuilder is the Future of Pickup and Delivery App Development (1).pdf
PDF
Design an Analysis of Algorithms I-SECS-1021-03
PDF
wealthsignaloriginal-com-DS-text-... (1).pdf
PPTX
assetexplorer- product-overview - presentation
PDF
PTS Company Brochure 2025 (1).pdf.......
PDF
Digital Systems & Binary Numbers (comprehensive )
PDF
Navsoft: AI-Powered Business Solutions & Custom Software Development
PPTX
Oracle E-Business Suite: A Comprehensive Guide for Modern Enterprises
PPTX
Why Generative AI is the Future of Content, Code & Creativity?
Addressing The Cult of Project Management Tools-Why Disconnected Work is Hold...
Adobe Illustrator 28.6 Crack My Vision of Vector Design
CHAPTER 2 - PM Management and IT Context
Design an Analysis of Algorithms II-SECS-1021-03
Nekopoi APK 2025 free lastest update
top salesforce developer skills in 2025.pdf
Designing Intelligence for the Shop Floor.pdf
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
Cost to Outsource Software Development in 2025
Which alternative to Crystal Reports is best for small or large businesses.pdf
iTop VPN Free 5.6.0.5262 Crack latest version 2025
Why TechBuilder is the Future of Pickup and Delivery App Development (1).pdf
Design an Analysis of Algorithms I-SECS-1021-03
wealthsignaloriginal-com-DS-text-... (1).pdf
assetexplorer- product-overview - presentation
PTS Company Brochure 2025 (1).pdf.......
Digital Systems & Binary Numbers (comprehensive )
Navsoft: AI-Powered Business Solutions & Custom Software Development
Oracle E-Business Suite: A Comprehensive Guide for Modern Enterprises
Why Generative AI is the Future of Content, Code & Creativity?

GlusterFs: a scalable file system for today's and tomorrow's big data

  • 1. Roberto Franchini franchini@celi.it Codemotion Milano 29/11/2014
  • 2. GlusterFS A scalable distributed file system
  • 3. whoami(1) 15 years of experience, proud to be a programmer Writes software for information extraction, nlp, opinion mining (@scale ), and a lot of other buzzwords Implements scalable architectures Plays with servers Member of the JUG-Torino coordination team franchini@celi.it http://www.celi.it http://www.blogmeter.it github.com/robfrank github.com/uim-celi twitter.com/robfrankie linkedin.com/in/robfrank
  • 4. The problem Identify a distributed and scalable file system for today's and tomorrow's Big Data
  • 5. Once upon a time 2008: One nfs share 1,5TB ought to be enough for anybody 2010: Herd of shares (1,5TB x N) ought to be enough for anybody Nobody couldn’t stop the data flood It was the time for something new
  • 6. Requirements Can be enlarged on demand No dedicated HW OS is preferred and trusted No specialized API No specialized Kernel POSIX compliance Zilions of big and small files No NAS or SAN (€€€€€)
  • 7. Clustered Scale-out General Purpose Storage Platform − POSIX-y Distributed File System − ...and so much more Built on commodity systems − x86_64 Linux ++ − POSIX filesystems underneath (XFS, EXT4) No central metadata Server (NO SPOF) Modular architecture for scale and functionality
  • 8. Common use cases Large Scale File Server Media / Content Distribution Network (CDN) Backup / Archive / Disaster Recovery (DR) High Performance Computing (HPC) Infrastructure as a Service (IaaS) storage layer Database offload (blobs) Unified Object Store + File Access
  • 9. Features ACL and Quota support Fault-tolerance Peer to peer Self-healing Fast setup up Enlarge on demand Shrink on demand Snapshot On premise phisical or virtual On cloud
  • 11. Architecture Peer / Node − cluster servers (glusterfs server) − Runs the gluster daemons and participates in volumes Brick − A filesystem mountpoint on servers − A unit of storage used as a capacity building block
  • 12. Bricks on a node
  • 13. Architecture Translator − Logic between bricks or subvolume that generate a subvolume with certain characteristic − distribute, replica, stripe are special translators to generate simil-RAID configuration − perfomance translators Volume − Bricks combined and passed through translators − Ultimately, what's presented to the end user
  • 16. Distributed The default configuration Files “evenly” spread across bricks Similar to file-level RAID 0 Server/Disk failure could be catastrophic
  • 18. Replicated Files written synchronously to replica peers Files read synchronously, but ultimately serviced by the first responder Similar to file-level RAID 1
  • 20. Distributed + replicated Distribued + replicated Similar to file-level RAID 10 Most used layout
  • 22. Striped Individual files split among bricks (sparse files) Similar to block-level RAID 0 Limited Use Cases HPC Pre/Post Processing File size exceeds brick size
  • 25. Components glusterd Management daemon One instance on each GlusterFS server Interfaced through gluster CLI glusterfsd GlusterFS brick daemon One process for each brick on each server Managed by glusterd
  • 26. Components glusterfs Volume service daemon One process for each volume service NFS server, FUSE client, Self-Heal, Quota, ... mount.glusterfs FUSE native client mount extension gluster Gluster Console Manager (CLI)
  • 28. Clients: native FUSE kernel module allows the filesystem to be built and operated entirely in userspace Specify mount to any GlusterFS server Native Client fetches volfile from mount server, then communicates directly with all nodes to access data Recommended for high concurrency and high write performance Load is inherently balanced across distributed volumes
  • 29. Clients:NFS Standard NFS v3 clients Standard automounter is supported Mount to any server, or use a load balancer GlusterFS NFS server includes Network Lock Manager (NLM) to synchronize locks across clients Better performance for reading many small files from a single client Load balancing must be managed externally
  • 30. Clients: libgfapi Introduced with GlusterFS 3.4 User-space library for accessing data in GlusterFS Filesystem-like API Runs in application process no FUSE, no copies, no context switches ...but same volfiles, translators, etc.
  • 31. Clients: SMB/CIFS In GlusterFS 3.4 – Samba + libgfapi No need for local native client mount & re-export Significant performance improvements with FUSE removed from the equation Must be setup on each server you wish to connect to via CIFS CTDB is required for Samba clustering
  • 32. Clients: HDFS Access data within and outside of Hadoop No HDFS name node single point of failure / bottleneck Seamless replacement for HDFS Scales with the massive growth of big data
  • 34. Under the hood Elastic Hash Algorithm No central metadata No Performance Bottleneck Eliminates risk scenarios Location hashed intelligently on filename Unique identifiers (GFID), similar to md5sum
  • 35. Scalability 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB Gluster Server 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB Gluster Server 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB 3TB Gluster Server Scale out performance and availability Scale out capacitry
  • 36. Scalability Add disks to servers to increase storage size Add servers to increase bandwidth and storage size Add servers to increase availability (replica factor)
  • 37. What we do with glusterFS
  • 38. What we do with GFS Daily production of more than 10GB of Lucene inverted indexes stored on glusterFS more than 200GB/month Search stored indexes to extract different sets of documents for every customers YES: we open indexes directly on storage (it's POSIX!!!)
  • 39. 2010: first installation Version 3.0.x 8 (not dedicated) servers Distributed replicated No bound on brick size (!!!!) Ca 4TB avaliable NOTE: stuck to 3.0.x until 2012 due to problems on 3.1 and 3.2 series, then RH acquired gluster (RH Storage)
  • 40. 2012: (little) cluster New installation, version 3.3.2 4TB available on 8 servers (DELL c5000) still not dedicated 1 brick per server limited to 1TB 2TB-raid 1 on each server Still in production
  • 41. 2012: enlarge New installation, upgrade to 3.3.x 6TB available on 12 servers (still not dedicated) Enlarged to 9TB on 18 servers Bricks size bounded AND unbounded
  • 42. 2013: fail 18 not dedicated servers: too much 18 bricks of different sizes 2 big down due to bricks out of space Didn’t restart after a move but… All data were recovered (files are scattered on bricks, read from them!)
  • 43. 2014: consolidate 2 dedicated servers 12 x 3TB SAS raid6 4 bricks per server 28 TB available distributed replicated 4x1Gb bonded NIC ca 40 clients (FUSE) (other servers)
  • 44. Consolidate brick 1 brick 2 brick 3 brick 4 Gluster Server 1 brick 1 brick 2 brick 3 brick 4 Gluster Server 2
  • 45. Scale up brick 11 brick 12 brick 13 brick 31 Gluster Server 1 brick 21 brick 22 brick 32 brick 24 Gluster Server 2 brick 31 brick 32 brick 23 brick 14 Gluster Server 3
  • 46. Do Dedicated server (phisical or virtual) RAID 6 or RAID 10 (with small files) Multiple bricks of same size Plan to scale
  • 47. Do not Multi purpose server Bricks of different size Very small files Write to bricks
  • 48. Some raw tests read Total transferred file size: 23.10G bytes 43.46M bytes/sec write Total transferred file size: 23.10G bytes 38.53M bytes/sec
  • 49. Raw tests NOTE: ran in production under heavy load, no clean test room
  • 50. Resources http://www.gluster.org/ https://access.redhat.com/documentation/en- US/Red_Hat_Storage/ https://github.com/gluster http://www.redhat.com/products/storage-server/ http://joejulian.name/blog/category/glusterfs/ http://jread.us/2013/06/one-petabyte-red-hat-storage-and-glusterfs- project-overview/