SlideShare a Scribd company logo
RED HAT CEPH STORAGE
ACCELERATION UTILIZING FLASH
TECHNOLOGY
Applications and Ecosystem Solutions Development
Rick Stehno
Red Hat Storage Day -
Boston 2016 1
Seagate Confidential 2
• Utilize flash caching features to accelerate critical data. Caching methods
can be write-back for writes, write-thru for disk/cache transparency, read
cache, etc..
• Utilize storage tiering capabilities. Performance critical data resides on
flash storage, colder data resides on HDD
• Utilize all flash storage to accelerate performance when all application
data is performance critical or when the application does not provide the
features or capabilities to cache or to migrate the data
Three ways to accelerate application performance with flash
Flash Acceleration for Applications
Seagate Confidential 3
Configurations:
• All flash storage - Performance
• Highest performance per node
• Less maximum capacity per node
• Hybrid HDD and flash storage - Balanced
• Balances performance, capacity and cost
• Application and workload suitable for
• Performance critical data on flash
• Utilize host software caching or tiering on flash
• All HDD storage - Capacity
• Maximum capacity per node, lowest cost
• Lower performance per node
Ceph Software Defined Storage (SDS) Acceleration
Seagate Confidential 4
–Higher performance in half the rack space
–28% less power and cooling
–Lower MTBF inherent with reduced component count
–Reduced OSD recovery time per Ceph node
–Lower TCO
Why 1U server with 10 NVMe SSDs may be better choice
vs. 2U Server with 24 SATA SSDs
Storage - NVMe vs SATA SSD
Seagate Confidential 5
• 4.5x increase for 128k sequential
reads
• 3.5x increase for 128k sequential
writes
• 3.7x increase for 4k random reads
• 1.4x increase for 4k random 70/30
RR/RW
• Equal performance for 4k random
writes
Why 1U server with 10 NVMe SSDs may be better choice
vs. 2U Server with 24 SATA SSDs
All Flash Storage - NVMe vs SATA SSD cont’d
FIO Benchmarks
Seagate Confidential 6
Why 1U server with 10 NVMe SSDs may be better choice
vs. 2U Server with 24 SATA SSDs
All Flash Storage - NVMe vs SATA SSD cont’d
Increasing the load to stress NVMe
capabilities over and above the 128
thread SATA SSD Test:
• 5.8x increase for Random Writes at
512 threads
• 3.1x increase for 70/30 RR/RW at
512 threads
• 4.2x increase for Random Reads at
790 threads
• 8.2x increase for Sequential Reads
at 1264 threads
3x
5.8x
1.4x
3.1x
1.0x
4.2x
1.3x
8.2x
128
Theads
512
Theads
128
Threads
512
Threads
128
threads
790
threads
128
threads
1264
threads
Gains
Random Write 70/30 RR/RW
Random Reads Sequential Reads
Ceph RBD NVMe Performance Gains over
SATA SSD
Random Writes 70/30 RR/RW Random Reads Sequential Reads
128k FIO RBD IOEngine Benchmark
Seagate Confidential 7
Price per MB/s: Cost of ((Retail Cost of SSD) / MB/s for each test)
SSD
Total SSD
Price
Price MB/s 128k Random Writes
128 threads
Price MB/s 128k Random Writes
512 threads
24 - SATA SSD 960G $7,896 24 - SATA SSD 960G $15.00
10 - NVMe 2TB $10,990 10 - NVMe 2TB $7.00 10 – NVMe 2TB $3.00
These prices do not include savings from electrical/cooling costs, reducing datacenter floor space, from the reduction of SATA SSD
Note: 128k random write FIO RBD benchmark: SATA SSD averaged 85% busy, NVMe averaged 80% busy with 512 threads
FIO RBD Maximum Threads Random Write Performance for NVMe
Ceph Storage Costs Seagate SATA SSD
vs. Seagate NVMe SSD
Seagate Confidential 8
MySQL
• MySQL is the most popular and the most widely used open-source database in the world
• MySQL is both feature rich in the areas of performance, scalability and reliability
• Database users demand high OLTP performance - Small random reads/writes
Ceph
• Most popular Software Defined Storage system
• Scalable
• Reliable
Does it make sense implementing Ceph into a MySQL
Database environment?
But Ceph was not designed to provide high performance for OLTP environments
OLTP entails small random reads/writes
Seagate Confidential 9
MySQL Setup:
Release 5.7
45,000,000 rows
6GB Buffer
4G logfiles
RAID 0 over 18 HDD
Ceph Setup:
3 Nodes each containing:
Jewel Using Filestore
4 NVMe SSDs
1 Pool over 12 NVMe SSDs
Replica 2
40G private and public
network
For all tests, all MySQL
files were local on local
server except the database
file, this file was moved to
the Ceph cluster.
MySQL - Comparing Local HDD to Ceph Cluster
Seagate Confidential 10Seagate Confidential
MySQL - Comparing Local HDD to Ceph Cluster
MySQL - Comparing Local HDD to Ceph Cluster
Seagate Confidential 11
MySQL - Comparing Local NVMe SSD to Ceph Cluster
MySQL Setup:
Release 5.7
45,000,000 rows
6GB Buffer
4G logfiles
RAID 0 over 4 NVMe SSDs
Ceph Setup:
3 Nodes each containing:
Jewel Using Filestore
4 NVMe SSDs
1 Pool over 12 NVMe SSDs
Replica 1
40G private and public
network
For all tests, all MySQL
files were local on local
server except the database
file, this file was moved to
the Ceph cluster.
Seagate Confidential 12
All SSD
Case-1: Case-2: Case-3:
2 SSDs 2 SSDs 1 PCIe flash
1 OSD/SSD 4 OSDs/SSD 4 OSDs/SSD
8 OSD journals on PCIe flash
0
100000
200000
300000
400000
500000
600000
700000
800000
0
200000
400000
600000
800000
1000000
1200000
2 ssd, 2 osd 2 ssd, 8 osd 2 ssd, 8 osd,
+journal
IOPS
KB/s
FIO Random Write - 200 Threads -
128k Data
Seagate SSD and Seagate PCIe Storage using AIC server
Ceph All Flash Storage Acceleration
Seagate Confidential 13
• Use RAW device or create 1st partition on 1M boundary (sector 2048)
• Ceph-deploy uses the optimal alignment when creating an OSD
• Use blk-mq/scsi-mq if kernel supports it
• rq_affinity = 1 for NVMe, rq_affinity = 2 for non-NVMe
• rotational = 0
• blockdev --setra 4096
Linux tuning is still a requirement to get optimum performance out of a SSD
Linux Flash Storage Tuning
Seagate Confidential 14
• If using an older kernel that doesn’t support BLK-MQ, use:
• “deadline” IO-Scheduler with supporting variables:
• fifo-batch
• front-merges
• writes-starved
• XFS Mount options:
• nobarrier,discard,noatime,attr2,delaylog,inode64,noquota
• If using a smaller number of SSD/NVMe SSD, test with creating multiple OSD’s per
SSD/NVMe SSD. Have seen good performance increases using 4 OSD per SSD/NVMe
SSD
• MySQL – when using flash, configure both innodb_io_capacity and
innodb_lru_scan_depth
Linux tuning is still a requirement to get optimum performance out of a SSD
Linux Flash Storage Tuning cont’d
Seagate Confidential 15
Flash Storage Device Configuration
If the NVMe SSD or SAS/SATA SSD device can be configured to use a 4k sector size,
this could increase performance for certain applications like databases.
For all of my FIO tests with the RBD engine and for all of my MySQL tests, I saw up to
a 2x improvement (depending on the test) when using 4k sector sizes compared to
using 512 byte sectors.
Storage devices used for all of the above benchmarks/tests:
• Seagate Nytro XF1440 NVMe SSD
• Seagate Nytro XF1230 SATA SSD
• Seagate 1200.2 SAS SSD
• Seagate XP6500 PCIe Flash Accelerator Card
Seagate Confidential 16
Seagate Broadest PCIe, SAS and SATA Portfolio
Seagate Confidential 17Seagate Confidential
Thank You!
Questions?
Learn how Seagate accelerates storage
with one of the broadest SSD and Flash
portfolios in the market

More Related Content

PDF
Red Hat Storage Day New York - New Reference Architectures
PPTX
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
PPTX
Ceph Deployment at Target: Customer Spotlight
PDF
Red Hat Storage Day New York - QCT: Avoid the mess, deploy with a validated s...
PPTX
Red Hat Storage Day Dallas - Defiance of the Appliance
PDF
Red Hat Storage Day New York -Performance Intensive Workloads with Samsung NV...
PDF
Red Hat Storage Day New York - Persistent Storage for Containers
PDF
Red Hat Storage Day New York - What's New in Red Hat Ceph Storage
Red Hat Storage Day New York - New Reference Architectures
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...
Ceph Deployment at Target: Customer Spotlight
Red Hat Storage Day New York - QCT: Avoid the mess, deploy with a validated s...
Red Hat Storage Day Dallas - Defiance of the Appliance
Red Hat Storage Day New York -Performance Intensive Workloads with Samsung NV...
Red Hat Storage Day New York - Persistent Storage for Containers
Red Hat Storage Day New York - What's New in Red Hat Ceph Storage

What's hot (19)

PDF
Red Hat Storage Day Boston - OpenStack + Ceph Storage
PPTX
Red Hat Storage Day Seattle: Stabilizing Petabyte Ceph Cluster in OpenStack C...
PDF
Build an High-Performance and High-Durable Block Storage Service Based on Ceph
PPTX
New Ceph capabilities and Reference Architectures
PPTX
Architecting Ceph Solutions
PPTX
MySQL Head-to-Head
PDF
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
PPTX
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
PPTX
Which Hypervisor is Best?
PDF
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
PPTX
MySQL Head to Head Performance
PDF
Storage tiering and erasure coding in Ceph (SCaLE13x)
PDF
Red Hat Storage for Mere Mortals
PPTX
QCT Ceph Solution - Design Consideration and Reference Architecture
PDF
Red Hat Storage Roadmap
PDF
Ceph as software define storage
PPTX
Designing for High Performance Ceph at Scale
PDF
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
PPTX
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Red Hat Storage Day Boston - OpenStack + Ceph Storage
Red Hat Storage Day Seattle: Stabilizing Petabyte Ceph Cluster in OpenStack C...
Build an High-Performance and High-Durable Block Storage Service Based on Ceph
New Ceph capabilities and Reference Architectures
Architecting Ceph Solutions
MySQL Head-to-Head
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
Which Hypervisor is Best?
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
MySQL Head to Head Performance
Storage tiering and erasure coding in Ceph (SCaLE13x)
Red Hat Storage for Mere Mortals
QCT Ceph Solution - Design Consideration and Reference Architecture
Red Hat Storage Roadmap
Ceph as software define storage
Designing for High Performance Ceph at Scale
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ad

Viewers also liked (19)

PDF
Red Hat Storage Day Boston - Persistent Storage for Containers
PPTX
Red Hat Storage Day Dallas - Why Software-defined Storage Matters
PDF
Red Hat Storage Day - When the Ceph Hits the Fan
PPTX
Red Hat Storage Day Dallas - Gluster Storage in Containerized Application
PPTX
Red Hat Storage Day Boston - Supermicro Super Storage
PDF
Red Hat Storage Day Dallas - Storage for OpenShift Containers
PPTX
Red Hat Storage Day Boston - Why Software-defined Storage Matters
PDF
Red Hat Storage Day New York - Red Hat Gluster Storage: Historical Tick Data ...
PDF
Red Hat Storage Day Boston - Red Hat Gluster Storage vs. Traditional Storage ...
PDF
Red Hat Storage Day New York - Penguin Computing Spotlight: Delivering Open S...
PDF
Red Hat Storage Day New York - Intel Unlocking Big Data Infrastructure Effici...
PDF
Red Hat Storage Day New York - Welcome Remarks
PPTX
Storage: Limitations, Frustrations, and Coping with Future Needs
PDF
CEPH introduction , Bootstrapping your first Ceph cluster in just 10 minutes
PDF
Red Hat Storage 2014 - Product(s) Overview
PDF
Red Hat Enterprise Linux OpenStack Platform on Inktank Ceph Enterprise
PDF
What's new in Jewel and Beyond
PPTX
My SQL on Ceph
PDF
Red Hat Ceph Storage Roadmap: January 2016
Red Hat Storage Day Boston - Persistent Storage for Containers
Red Hat Storage Day Dallas - Why Software-defined Storage Matters
Red Hat Storage Day - When the Ceph Hits the Fan
Red Hat Storage Day Dallas - Gluster Storage in Containerized Application
Red Hat Storage Day Boston - Supermicro Super Storage
Red Hat Storage Day Dallas - Storage for OpenShift Containers
Red Hat Storage Day Boston - Why Software-defined Storage Matters
Red Hat Storage Day New York - Red Hat Gluster Storage: Historical Tick Data ...
Red Hat Storage Day Boston - Red Hat Gluster Storage vs. Traditional Storage ...
Red Hat Storage Day New York - Penguin Computing Spotlight: Delivering Open S...
Red Hat Storage Day New York - Intel Unlocking Big Data Infrastructure Effici...
Red Hat Storage Day New York - Welcome Remarks
Storage: Limitations, Frustrations, and Coping with Future Needs
CEPH introduction , Bootstrapping your first Ceph cluster in just 10 minutes
Red Hat Storage 2014 - Product(s) Overview
Red Hat Enterprise Linux OpenStack Platform on Inktank Ceph Enterprise
What's new in Jewel and Beyond
My SQL on Ceph
Red Hat Ceph Storage Roadmap: January 2016
Ad

Similar to Red Hat Ceph Storage Acceleration Utilizing Flash Technology (20)

PPTX
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
PPTX
Implementation of Dense Storage Utilizing HDDs with SSDs and PCIe Flash Acc...
PDF
Seagate Implementation of Dense Storage Utilizing HDDs and SSDs
PPTX
Ceph Performance and Sizing Guide
PDF
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
PDF
Ceph Day Beijing - Ceph all-flash array design based on NUMA architecture
PPTX
openSUSE storage workshop 2016
PDF
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
PDF
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
PPTX
Ceph Community Talk on High-Performance Solid Sate Ceph
PPTX
Red Hat Storage Day LA - Performance and Sizing Software Defined Storage
PDF
Red hat Storage Day LA - Designing Ceph Clusters Using Intel-Based Hardware
PPTX
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
PPTX
Ceph Performance Profiling and Reporting
ODP
Exploiting Your File System to Build Robust & Efficient Workflows
PDF
Ceph Day Taipei - Bring Ceph to Enterprise
PPTX
Ceph Day Tokyo - Bring Ceph to Enterprise
PDF
NAVER Ceph Storage on ssd for Container
PDF
Why does my choice of storage matter with cassandra?
PDF
Database performance tuning for SSD based storage
Ceph Day San Jose - Red Hat Storage Acceleration Utlizing Flash Technology
Implementation of Dense Storage Utilizing HDDs with SSDs and PCIe Flash Acc...
Seagate Implementation of Dense Storage Utilizing HDDs and SSDs
Ceph Performance and Sizing Guide
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA Architecture
Ceph Day Beijing - Ceph all-flash array design based on NUMA architecture
openSUSE storage workshop 2016
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Ceph Community Talk on High-Performance Solid Sate Ceph
Red Hat Storage Day LA - Performance and Sizing Software Defined Storage
Red hat Storage Day LA - Designing Ceph Clusters Using Intel-Based Hardware
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...
Ceph Performance Profiling and Reporting
Exploiting Your File System to Build Robust & Efficient Workflows
Ceph Day Taipei - Bring Ceph to Enterprise
Ceph Day Tokyo - Bring Ceph to Enterprise
NAVER Ceph Storage on ssd for Container
Why does my choice of storage matter with cassandra?
Database performance tuning for SSD based storage

More from Red_Hat_Storage (7)

PPTX
Red Hat Storage Day Seattle: Why Software-Defined Storage Matters
PPTX
Red Hat Storage Day Seattle: Persistent Storage for Containerized Applications
PPTX
Red Hat Storage Day Seattle: Stretching A Gluster Cluster for Resilient Messa...
PDF
Red Hat Storage Day Atlanta - Persistent Storage for Linux Containers
PDF
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
PPTX
Red Hat Storage Day Atlanta - Why Software Defined Storage Matters
PPTX
Red Hat Storage Day Atlanta - Red Hat Gluster Storage vs. Traditional Storage...
Red Hat Storage Day Seattle: Why Software-Defined Storage Matters
Red Hat Storage Day Seattle: Persistent Storage for Containerized Applications
Red Hat Storage Day Seattle: Stretching A Gluster Cluster for Resilient Messa...
Red Hat Storage Day Atlanta - Persistent Storage for Linux Containers
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...
Red Hat Storage Day Atlanta - Why Software Defined Storage Matters
Red Hat Storage Day Atlanta - Red Hat Gluster Storage vs. Traditional Storage...

Recently uploaded (20)

PDF
GamePlan Trading System Review: Professional Trader's Honest Take
PDF
Modernizing your data center with Dell and AMD
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
Spectral efficient network and resource selection model in 5G networks
PPTX
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PPTX
Big Data Technologies - Introduction.pptx
PDF
GDG Cloud Iasi [PUBLIC] Florian Blaga - Unveiling the Evolution of Cybersecur...
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PDF
Machine learning based COVID-19 study performance prediction
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
Advanced IT Governance
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
GamePlan Trading System Review: Professional Trader's Honest Take
Modernizing your data center with Dell and AMD
Diabetes mellitus diagnosis method based random forest with bat algorithm
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Per capita expenditure prediction using model stacking based on satellite ima...
Review of recent advances in non-invasive hemoglobin estimation
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Spectral efficient network and resource selection model in 5G networks
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
The Rise and Fall of 3GPP – Time for a Sabbatical?
Big Data Technologies - Introduction.pptx
GDG Cloud Iasi [PUBLIC] Florian Blaga - Unveiling the Evolution of Cybersecur...
Advanced methodologies resolving dimensionality complications for autism neur...
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
Machine learning based COVID-19 study performance prediction
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Advanced IT Governance
Mobile App Security Testing_ A Comprehensive Guide.pdf

Red Hat Ceph Storage Acceleration Utilizing Flash Technology

  • 1. RED HAT CEPH STORAGE ACCELERATION UTILIZING FLASH TECHNOLOGY Applications and Ecosystem Solutions Development Rick Stehno Red Hat Storage Day - Boston 2016 1
  • 2. Seagate Confidential 2 • Utilize flash caching features to accelerate critical data. Caching methods can be write-back for writes, write-thru for disk/cache transparency, read cache, etc.. • Utilize storage tiering capabilities. Performance critical data resides on flash storage, colder data resides on HDD • Utilize all flash storage to accelerate performance when all application data is performance critical or when the application does not provide the features or capabilities to cache or to migrate the data Three ways to accelerate application performance with flash Flash Acceleration for Applications
  • 3. Seagate Confidential 3 Configurations: • All flash storage - Performance • Highest performance per node • Less maximum capacity per node • Hybrid HDD and flash storage - Balanced • Balances performance, capacity and cost • Application and workload suitable for • Performance critical data on flash • Utilize host software caching or tiering on flash • All HDD storage - Capacity • Maximum capacity per node, lowest cost • Lower performance per node Ceph Software Defined Storage (SDS) Acceleration
  • 4. Seagate Confidential 4 –Higher performance in half the rack space –28% less power and cooling –Lower MTBF inherent with reduced component count –Reduced OSD recovery time per Ceph node –Lower TCO Why 1U server with 10 NVMe SSDs may be better choice vs. 2U Server with 24 SATA SSDs Storage - NVMe vs SATA SSD
  • 5. Seagate Confidential 5 • 4.5x increase for 128k sequential reads • 3.5x increase for 128k sequential writes • 3.7x increase for 4k random reads • 1.4x increase for 4k random 70/30 RR/RW • Equal performance for 4k random writes Why 1U server with 10 NVMe SSDs may be better choice vs. 2U Server with 24 SATA SSDs All Flash Storage - NVMe vs SATA SSD cont’d FIO Benchmarks
  • 6. Seagate Confidential 6 Why 1U server with 10 NVMe SSDs may be better choice vs. 2U Server with 24 SATA SSDs All Flash Storage - NVMe vs SATA SSD cont’d Increasing the load to stress NVMe capabilities over and above the 128 thread SATA SSD Test: • 5.8x increase for Random Writes at 512 threads • 3.1x increase for 70/30 RR/RW at 512 threads • 4.2x increase for Random Reads at 790 threads • 8.2x increase for Sequential Reads at 1264 threads 3x 5.8x 1.4x 3.1x 1.0x 4.2x 1.3x 8.2x 128 Theads 512 Theads 128 Threads 512 Threads 128 threads 790 threads 128 threads 1264 threads Gains Random Write 70/30 RR/RW Random Reads Sequential Reads Ceph RBD NVMe Performance Gains over SATA SSD Random Writes 70/30 RR/RW Random Reads Sequential Reads 128k FIO RBD IOEngine Benchmark
  • 7. Seagate Confidential 7 Price per MB/s: Cost of ((Retail Cost of SSD) / MB/s for each test) SSD Total SSD Price Price MB/s 128k Random Writes 128 threads Price MB/s 128k Random Writes 512 threads 24 - SATA SSD 960G $7,896 24 - SATA SSD 960G $15.00 10 - NVMe 2TB $10,990 10 - NVMe 2TB $7.00 10 – NVMe 2TB $3.00 These prices do not include savings from electrical/cooling costs, reducing datacenter floor space, from the reduction of SATA SSD Note: 128k random write FIO RBD benchmark: SATA SSD averaged 85% busy, NVMe averaged 80% busy with 512 threads FIO RBD Maximum Threads Random Write Performance for NVMe Ceph Storage Costs Seagate SATA SSD vs. Seagate NVMe SSD
  • 8. Seagate Confidential 8 MySQL • MySQL is the most popular and the most widely used open-source database in the world • MySQL is both feature rich in the areas of performance, scalability and reliability • Database users demand high OLTP performance - Small random reads/writes Ceph • Most popular Software Defined Storage system • Scalable • Reliable Does it make sense implementing Ceph into a MySQL Database environment? But Ceph was not designed to provide high performance for OLTP environments OLTP entails small random reads/writes
  • 9. Seagate Confidential 9 MySQL Setup: Release 5.7 45,000,000 rows 6GB Buffer 4G logfiles RAID 0 over 18 HDD Ceph Setup: 3 Nodes each containing: Jewel Using Filestore 4 NVMe SSDs 1 Pool over 12 NVMe SSDs Replica 2 40G private and public network For all tests, all MySQL files were local on local server except the database file, this file was moved to the Ceph cluster. MySQL - Comparing Local HDD to Ceph Cluster
  • 10. Seagate Confidential 10Seagate Confidential MySQL - Comparing Local HDD to Ceph Cluster MySQL - Comparing Local HDD to Ceph Cluster
  • 11. Seagate Confidential 11 MySQL - Comparing Local NVMe SSD to Ceph Cluster MySQL Setup: Release 5.7 45,000,000 rows 6GB Buffer 4G logfiles RAID 0 over 4 NVMe SSDs Ceph Setup: 3 Nodes each containing: Jewel Using Filestore 4 NVMe SSDs 1 Pool over 12 NVMe SSDs Replica 1 40G private and public network For all tests, all MySQL files were local on local server except the database file, this file was moved to the Ceph cluster.
  • 12. Seagate Confidential 12 All SSD Case-1: Case-2: Case-3: 2 SSDs 2 SSDs 1 PCIe flash 1 OSD/SSD 4 OSDs/SSD 4 OSDs/SSD 8 OSD journals on PCIe flash 0 100000 200000 300000 400000 500000 600000 700000 800000 0 200000 400000 600000 800000 1000000 1200000 2 ssd, 2 osd 2 ssd, 8 osd 2 ssd, 8 osd, +journal IOPS KB/s FIO Random Write - 200 Threads - 128k Data Seagate SSD and Seagate PCIe Storage using AIC server Ceph All Flash Storage Acceleration
  • 13. Seagate Confidential 13 • Use RAW device or create 1st partition on 1M boundary (sector 2048) • Ceph-deploy uses the optimal alignment when creating an OSD • Use blk-mq/scsi-mq if kernel supports it • rq_affinity = 1 for NVMe, rq_affinity = 2 for non-NVMe • rotational = 0 • blockdev --setra 4096 Linux tuning is still a requirement to get optimum performance out of a SSD Linux Flash Storage Tuning
  • 14. Seagate Confidential 14 • If using an older kernel that doesn’t support BLK-MQ, use: • “deadline” IO-Scheduler with supporting variables: • fifo-batch • front-merges • writes-starved • XFS Mount options: • nobarrier,discard,noatime,attr2,delaylog,inode64,noquota • If using a smaller number of SSD/NVMe SSD, test with creating multiple OSD’s per SSD/NVMe SSD. Have seen good performance increases using 4 OSD per SSD/NVMe SSD • MySQL – when using flash, configure both innodb_io_capacity and innodb_lru_scan_depth Linux tuning is still a requirement to get optimum performance out of a SSD Linux Flash Storage Tuning cont’d
  • 15. Seagate Confidential 15 Flash Storage Device Configuration If the NVMe SSD or SAS/SATA SSD device can be configured to use a 4k sector size, this could increase performance for certain applications like databases. For all of my FIO tests with the RBD engine and for all of my MySQL tests, I saw up to a 2x improvement (depending on the test) when using 4k sector sizes compared to using 512 byte sectors. Storage devices used for all of the above benchmarks/tests: • Seagate Nytro XF1440 NVMe SSD • Seagate Nytro XF1230 SATA SSD • Seagate 1200.2 SAS SSD • Seagate XP6500 PCIe Flash Accelerator Card
  • 16. Seagate Confidential 16 Seagate Broadest PCIe, SAS and SATA Portfolio
  • 17. Seagate Confidential 17Seagate Confidential Thank You! Questions? Learn how Seagate accelerates storage with one of the broadest SSD and Flash portfolios in the market

Editor's Notes

  • #4: SMRs – Drive Managed
  • #8: SMRs – Drive Managed
  • #13: SMRs – Drive Managed