SlideShare a Scribd company logo
Managing the data explosion
                      The causes, effects and solutions of growing data volumes

By Julian Stuhler
Director, Triton Consulting & IDUG President

Drowning in data
As Information Technology becomes ever more prevalent in nearly every aspect of our lives, the
amount of data generated and stored continues to grow at an astounding rate. According to IBM,
worldwide data volumes are currently doubling every two years. IDC estimates that 45GB of data
currently exists for each person on the planet: that’s a mind-blowing 281 Billion Gigabytes (281
Exabytes) in total. While a mere 5% of that data will end up on Enterprise data servers, it is forecast
to grow at a staggering 60% per year, resulting in 14 Exabytes of corporate data by 2011.

This article takes a look at some of the reasons behind this data explosion, and some of the possible
effects if the growth is not managed. We’ll also examine some of the ways in which these problems
can be avoided.

Industry Trends
A major trend over the last few years has seen many organisations implementing ERP and CRM
solutions. This in turn has caused a dramatic increase in the amount of data we are storing about
our customers, prospects, partners and suppliers.

Companies are also investing in ever more sophisticated business intelligence and analytics. In an
increasingly competitive marketplace, the ability to base business decisions on solid, reliable and
timely management information is becoming a key differentiator, but trend analysis can require very
large amounts of historical data to be stored and managed.

The trend towards company consolidation is not a new one, but the current economic situation has
inevitably resulted in a significant increase in the number of mergers and acquisitions. This is
creating a huge increase in data volumes, with the associated data duplication and application
retirement issues. Organisations are faced with not only managing all of their own data, both
historic and current but this influx of additional data from other parties. Imagine the “data
headache” of combining all of the ERP, CRM, Business Intelligence and Analytic systems from
different organisations into one manageable enterprise system.

Legislation
Corporate compliance legislation has had a major effect on how we use, store and maintain our
data. The requirements placed on organisations by HIPPA, SOX, Basel ll and others mean that many
companies are having to keep hold of more data, and for longer periods. Just as importantly, that
retained data rapidly transforms from a corporate asset to a liability once the legal minimum
retention period has expired, making it vital that such data can be accurately identified and deleted.
It is vital that organisations adhere to this legislation in order to avoid the cost of court appearances,
heavy fines and the resultant damage to the brand.

Technical Trends
New capabilities within the databases used to store corporate information are another major driver
of data growth. For example, DB2 now supports XML and LOBs (“large objects” such as audio, video,
images, etc). The ability to store this kind of data alongside more traditional structured information
can be very useful, but can also have a huge impact on the overall size of the database.
Other technical trends that are contributing to database growth include storage of data in Unicode
format (which can often expand overall database size by 10%-50% depending on the data), and
duplication of databases due to replication requirements and/or backup strategies.

Finally, there’s the perennial problem of removing old or obsolete data once it has reached the end
of its useful life. Application data archiving is often considered as an optional extra, and even if it is
included in the initial project plan it is often the first item to be postponed to a later release.

Effects of rapid data growth
This unprecedented growth in data volumes is having a significant effect on many organisations.
Perhaps the most obvious impact is on operational costs. More staff time is required for routine
maintenance and data-related exception handling such as out-of-space conditions and re-
partitioning. As the database increases in size, so too does the CPU cost of running batch operations
and routine housekeeping. Ongoing running costs also increase due to the additional disk space
required and storage and processing capacity upgrades may be needed even though they often
haven’t been budgeted for.

Painful though they may be, increases in operational costs aren’t the end of the story. What price
can you place on customer satisfaction? Performance for critical application processes can degrade
as data volumes increase, resulting in missed service level objectives. Teams across the whole
organisation may be affected, with call centre staff unable to access the information they need
quickly enough to satisfy customer demand.

Coping with the data explosion
Various coping strategies are available to address the issues associated with rapid data growth.
Measures such as implementing database partitioning and data compression or purchasing extra
CPU/DASD can help. However, these have their own costs there are many issues still remaining,
including:
            – Disaster recovery times
            – Legal risk of exceeding minimum data retention periods (data as a liability, not an
                asset)
            – DBA effort to manage/tune workloads and databases
            – Cost of spending IT budget on maintaining current capacity, not innovating


So, what are the alternatives?
Implement a data archiving strategy!

According to a recent Gartner report “database archiving significantly lowers storage costs for
primary storage by moving older data to less-costly storage” they go on to say “archiving reduces the
size of primary storage, resulting in improved application performance and lower storage
requirements for copies of the database for testing, backup and other purposes”

Also, you may think that archiving is only applicable to the largest of applications but in the same
report Gartner state that “Performance and cost improvements can be sizeable, even with
applications that have less than 200GB of data”

So, it would appear that a data archiving strategy is the best way for organisations to cope with
growing data. Giving cost savings and improved application performance. However, once the need
to archive has been agreed many new questions arise:
•   Build Vs Buy
    •   Flexibility Vs Speed
    •   Software expenditure Vs staff time costs

These are the tough decisions which need to be made before a data archiving strategy can be put
into place. While the temptation to build in house may be strong is there really justification for so
doing? Can staff be spared to work on this project? Although the up-front cost is cheaper what
about the long-term cost, not just in staff time for the project but ongoing as expertise is lost
through staff movement. What about the need to implement the strategy across multiple platforms
within the same organisation? Can we spare project staff from each area of the organisation to
work on developing a bespoke solution for their operating platform?

The answer is potentially a bought in solution which will work across multiple platforms thus
bringing a scalable solution to the enterprise without needing to take precious staff time away onto
separate, long-term test and development projects to create a bespoke solution.

So, it seems that there are ways to control data growth before it controls us. By implementing a
thorough archiving policy and an “intelligent archiving” system we can manage data throughout it’s
lifecycle.



Triton Consulting are Information Management Specialists and IBM Premier Business Partners. For
more information on Triton and the solutions they provide visit www.triton.co.uk

www.bloor-research.com

More Related Content

PDF
EMC Isilon: A Scalable Storage Platform for Big Data
 
PDF
TierPoint_ColocationWhitepaper-Six_Reasons
PDF
Big Data Management: Work Smarter Not Harder
PDF
Hadoop: Data Storage Locker or Agile Analytics Platform? It’s Up to You.
PDF
Move It Don't Lose It: Is Your Big Data Collecting Dust?
PDF
Kusnetzky Group: Maximizing the Benefits of Virtualization with Real-time Co...
PDF
Database Archiving: The Key to Siebel Performance
PDF
Building Smarter, Faster, and Scalable Data-Rich Application
EMC Isilon: A Scalable Storage Platform for Big Data
 
TierPoint_ColocationWhitepaper-Six_Reasons
Big Data Management: Work Smarter Not Harder
Hadoop: Data Storage Locker or Agile Analytics Platform? It’s Up to You.
Move It Don't Lose It: Is Your Big Data Collecting Dust?
Kusnetzky Group: Maximizing the Benefits of Virtualization with Real-time Co...
Database Archiving: The Key to Siebel Performance
Building Smarter, Faster, and Scalable Data-Rich Application

What's hot (19)

PDF
Enabling Cloud Analytics with Data-Level Security
PDF
BRIDGING DATA SILOS USING BIG DATA INTEGRATION
PDF
Maximizing the Benefits of Virtualization with Real-­time Compression
PDF
Matthew Johnston - Big Data Futures Outlook BCM
PDF
Big Data Analytics
PDF
The ABCs of Big Data
PDF
Data Management
PDF
Analytics3.0 e book
PDF
Trends 2011 and_beyond_business_intelligence
PDF
Creating the Foundations for the Internet of Things
PDF
Practical analytics john enoch white paper
PDF
Hadoop Overview
PDF
Big data-analytics-2013-peer-research-report
PDF
Dealing with Dark Data
PDF
Why You Need to Govern Big Data
PPT
Tesco data centre consolidation ‘nearing completion’
PDF
Sgcp14dunlea
PPTX
Week4 Day4
Enabling Cloud Analytics with Data-Level Security
BRIDGING DATA SILOS USING BIG DATA INTEGRATION
Maximizing the Benefits of Virtualization with Real-­time Compression
Matthew Johnston - Big Data Futures Outlook BCM
Big Data Analytics
The ABCs of Big Data
Data Management
Analytics3.0 e book
Trends 2011 and_beyond_business_intelligence
Creating the Foundations for the Internet of Things
Practical analytics john enoch white paper
Hadoop Overview
Big data-analytics-2013-peer-research-report
Dealing with Dark Data
Why You Need to Govern Big Data
Tesco data centre consolidation ‘nearing completion’
Sgcp14dunlea
Week4 Day4
Ad

Similar to Managing The Data Explosion (20)

PDF
Case Studies in Improving Application Performance With Solix Database Archivi...
PDF
Delphix_IDC_Analyst_Report_Holistic.pdf-aliId=496034
DOCX
How Analytics Has Changed in the Last 10 Years (and How It’s Staye.docx
PDF
A&D In Memory POV R2.2
PDF
Business Intelligence Solution on Windows Azure
PDF
The New Enterprise Blueprint featuring the Gartner Magic Quadrant
PDF
Six Reasons to Upgrade your Database
PDF
How companies are managing growth, gaining insights and cutting costs in the ...
PDF
Six Reasons to Upgrade your Database
PDF
IBM 2016 - Six reasons to upgrade your database
PDF
Activating Big Data: The Key To Success with Machine Learning Advanced Analyt...
PDF
Big Data is Here for Financial Services White Paper
PDF
9 Steps to Successful Information Lifecycle Management
PDF
Creating a Successful DataOps Framework for Your Business.pdf
PDF
Solix Cloud – Managing Data Growth with Database Archiving and Application Re...
PDF
Hitachi white-paper-storage-virtualization
PDF
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
PDF
Hitachi white-paper-future-proof-your-datacenter-with-the-right-nas-platform
PDF
Mighty Guides- Data Disruption
PDF
Ab cs of big data
Case Studies in Improving Application Performance With Solix Database Archivi...
Delphix_IDC_Analyst_Report_Holistic.pdf-aliId=496034
How Analytics Has Changed in the Last 10 Years (and How It’s Staye.docx
A&D In Memory POV R2.2
Business Intelligence Solution on Windows Azure
The New Enterprise Blueprint featuring the Gartner Magic Quadrant
Six Reasons to Upgrade your Database
How companies are managing growth, gaining insights and cutting costs in the ...
Six Reasons to Upgrade your Database
IBM 2016 - Six reasons to upgrade your database
Activating Big Data: The Key To Success with Machine Learning Advanced Analyt...
Big Data is Here for Financial Services White Paper
9 Steps to Successful Information Lifecycle Management
Creating a Successful DataOps Framework for Your Business.pdf
Solix Cloud – Managing Data Growth with Database Archiving and Application Re...
Hitachi white-paper-storage-virtualization
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
Hitachi white-paper-future-proof-your-datacenter-with-the-right-nas-platform
Mighty Guides- Data Disruption
Ab cs of big data
Ad

More from Laura Hood (20)

PDF
Top 10 DB2 Support Nightmares #10
PDF
Top 10 DB2 Support Nightmares #9
PDF
Top 10 DB2 Support Nightmares #8
PDF
Top 10 DB2 Support Nightmares #7
PDF
Top 10 db2 support nightmares #6
PDF
Consultancy on Demand - Infographic
PDF
A Time Traveller's Guide to DB2: Technology Themes for 2014 and Beyond
PDF
Top 10 DB2 Support Nightmares #1
PDF
Db2 10 memory management uk db2 user group june 2013 [read-only]
PDF
DB2 10 Security Enhancements
PDF
DbB 10 Webcast #3 The Secrets Of Scalability
PDF
DB2 10 Webcast #2 - Justifying The Upgrade
PDF
DB2 10 Webcast #1 - Overview And Migration Planning
PDF
Time Travelling With DB2 10 For zOS
PDF
DB2DART - DB2Night Show October 2011
PDF
DB2 z/OS & Java - What\'s New?
PDF
Temporal And Other DB2 10 For Z Os Highlights
PDF
DB210 Smarter Database IBM Tech Forum 2011
PDF
UKGSE DB2 pureScale
PPTX
UKCMG DB2 pureScale
Top 10 DB2 Support Nightmares #10
Top 10 DB2 Support Nightmares #9
Top 10 DB2 Support Nightmares #8
Top 10 DB2 Support Nightmares #7
Top 10 db2 support nightmares #6
Consultancy on Demand - Infographic
A Time Traveller's Guide to DB2: Technology Themes for 2014 and Beyond
Top 10 DB2 Support Nightmares #1
Db2 10 memory management uk db2 user group june 2013 [read-only]
DB2 10 Security Enhancements
DbB 10 Webcast #3 The Secrets Of Scalability
DB2 10 Webcast #2 - Justifying The Upgrade
DB2 10 Webcast #1 - Overview And Migration Planning
Time Travelling With DB2 10 For zOS
DB2DART - DB2Night Show October 2011
DB2 z/OS & Java - What\'s New?
Temporal And Other DB2 10 For Z Os Highlights
DB210 Smarter Database IBM Tech Forum 2011
UKGSE DB2 pureScale
UKCMG DB2 pureScale

Managing The Data Explosion

  • 1. Managing the data explosion The causes, effects and solutions of growing data volumes By Julian Stuhler Director, Triton Consulting & IDUG President Drowning in data As Information Technology becomes ever more prevalent in nearly every aspect of our lives, the amount of data generated and stored continues to grow at an astounding rate. According to IBM, worldwide data volumes are currently doubling every two years. IDC estimates that 45GB of data currently exists for each person on the planet: that’s a mind-blowing 281 Billion Gigabytes (281 Exabytes) in total. While a mere 5% of that data will end up on Enterprise data servers, it is forecast to grow at a staggering 60% per year, resulting in 14 Exabytes of corporate data by 2011. This article takes a look at some of the reasons behind this data explosion, and some of the possible effects if the growth is not managed. We’ll also examine some of the ways in which these problems can be avoided. Industry Trends A major trend over the last few years has seen many organisations implementing ERP and CRM solutions. This in turn has caused a dramatic increase in the amount of data we are storing about our customers, prospects, partners and suppliers. Companies are also investing in ever more sophisticated business intelligence and analytics. In an increasingly competitive marketplace, the ability to base business decisions on solid, reliable and timely management information is becoming a key differentiator, but trend analysis can require very large amounts of historical data to be stored and managed. The trend towards company consolidation is not a new one, but the current economic situation has inevitably resulted in a significant increase in the number of mergers and acquisitions. This is creating a huge increase in data volumes, with the associated data duplication and application retirement issues. Organisations are faced with not only managing all of their own data, both historic and current but this influx of additional data from other parties. Imagine the “data headache” of combining all of the ERP, CRM, Business Intelligence and Analytic systems from different organisations into one manageable enterprise system. Legislation Corporate compliance legislation has had a major effect on how we use, store and maintain our data. The requirements placed on organisations by HIPPA, SOX, Basel ll and others mean that many companies are having to keep hold of more data, and for longer periods. Just as importantly, that retained data rapidly transforms from a corporate asset to a liability once the legal minimum retention period has expired, making it vital that such data can be accurately identified and deleted. It is vital that organisations adhere to this legislation in order to avoid the cost of court appearances, heavy fines and the resultant damage to the brand. Technical Trends New capabilities within the databases used to store corporate information are another major driver of data growth. For example, DB2 now supports XML and LOBs (“large objects” such as audio, video, images, etc). The ability to store this kind of data alongside more traditional structured information can be very useful, but can also have a huge impact on the overall size of the database.
  • 2. Other technical trends that are contributing to database growth include storage of data in Unicode format (which can often expand overall database size by 10%-50% depending on the data), and duplication of databases due to replication requirements and/or backup strategies. Finally, there’s the perennial problem of removing old or obsolete data once it has reached the end of its useful life. Application data archiving is often considered as an optional extra, and even if it is included in the initial project plan it is often the first item to be postponed to a later release. Effects of rapid data growth This unprecedented growth in data volumes is having a significant effect on many organisations. Perhaps the most obvious impact is on operational costs. More staff time is required for routine maintenance and data-related exception handling such as out-of-space conditions and re- partitioning. As the database increases in size, so too does the CPU cost of running batch operations and routine housekeeping. Ongoing running costs also increase due to the additional disk space required and storage and processing capacity upgrades may be needed even though they often haven’t been budgeted for. Painful though they may be, increases in operational costs aren’t the end of the story. What price can you place on customer satisfaction? Performance for critical application processes can degrade as data volumes increase, resulting in missed service level objectives. Teams across the whole organisation may be affected, with call centre staff unable to access the information they need quickly enough to satisfy customer demand. Coping with the data explosion Various coping strategies are available to address the issues associated with rapid data growth. Measures such as implementing database partitioning and data compression or purchasing extra CPU/DASD can help. However, these have their own costs there are many issues still remaining, including: – Disaster recovery times – Legal risk of exceeding minimum data retention periods (data as a liability, not an asset) – DBA effort to manage/tune workloads and databases – Cost of spending IT budget on maintaining current capacity, not innovating So, what are the alternatives? Implement a data archiving strategy! According to a recent Gartner report “database archiving significantly lowers storage costs for primary storage by moving older data to less-costly storage” they go on to say “archiving reduces the size of primary storage, resulting in improved application performance and lower storage requirements for copies of the database for testing, backup and other purposes” Also, you may think that archiving is only applicable to the largest of applications but in the same report Gartner state that “Performance and cost improvements can be sizeable, even with applications that have less than 200GB of data” So, it would appear that a data archiving strategy is the best way for organisations to cope with growing data. Giving cost savings and improved application performance. However, once the need to archive has been agreed many new questions arise:
  • 3. Build Vs Buy • Flexibility Vs Speed • Software expenditure Vs staff time costs These are the tough decisions which need to be made before a data archiving strategy can be put into place. While the temptation to build in house may be strong is there really justification for so doing? Can staff be spared to work on this project? Although the up-front cost is cheaper what about the long-term cost, not just in staff time for the project but ongoing as expertise is lost through staff movement. What about the need to implement the strategy across multiple platforms within the same organisation? Can we spare project staff from each area of the organisation to work on developing a bespoke solution for their operating platform? The answer is potentially a bought in solution which will work across multiple platforms thus bringing a scalable solution to the enterprise without needing to take precious staff time away onto separate, long-term test and development projects to create a bespoke solution. So, it seems that there are ways to control data growth before it controls us. By implementing a thorough archiving policy and an “intelligent archiving” system we can manage data throughout it’s lifecycle. Triton Consulting are Information Management Specialists and IBM Premier Business Partners. For more information on Triton and the solutions they provide visit www.triton.co.uk www.bloor-research.com