The Data Center Journal recently featured an article I wrote entitled "Extending Encryption."
In this article, I write that data breaches are an everyday occurrence, and the consequences for organizations that suffer a breach can be horrendous, ranging from loss of consumer confidence to fines and punitive damages.
Some organizations, however, have discovered through experience that there is one tool that can help: encryption. For instance, last fall, the Second Appellate District Court of California ruled in favor of the defendant, the University of California, in a breach dating back to 2011 involving the theft of a laptop with 16,000 patient records. A plaintiff had sued under the Confidentiality of Medical Information Act (CMIA), a state statute. The case was decided in part on the fact that the patient records on the laptop were encrypted, thus nullifying any notion of actual loss of confidentiality or privacy.
To be sure, encryption of data “in flight” is increasingly the norm, and many organizations now also recognize the importance of encrypting data “at rest” on a storage device. But despite the growing need to protect sensitive data to meet regulatory and compliance mandates as well as service-level agreements, many are hesitant to make a wide-ranging embrace of the technology. The resistance to enabling encryption of data at rest has to do with a number of important factors: complexity, risk and costs. From a complexity standpoint, the problem is having to deal with many key managers and applications, most of which have proprietary key managers, along with the day-to-day management of multiple encryption-key-management (EKM) systems. Risks and cost go hand in hand: the risk of not meeting a backup window certainly presents heavy risk on any enterprise with the goal to meet and improve on service levels. Now imagine the resistance to enabling encryption of data at rest given that doing so on most systems, will increase the time to safety of the data on the backup appliance, thus increasing the risk. From a cost perspective, imagine the time lost by having to manage all these disparate systems daily versus a centralized key-management approach to solving this issue, as well as a single data-protection platform to manage.
Still, few can ignore the worsening threat that is bedeviling security professionals. Clearly, data inside the firewall can no longer be considered safe. Furthermore, the rise of cloud data storage and cloud computing in general means organizations must now vouch for data in environments they don’t even directly control. Those trends are now driving demand for and interest in protecting data at rest via encryption.
Fortunately, a double revolution is brewing. First, complexity associated with managing encryption keys is no longer as great a barrier thanks to developments such as the Key Management Interoperability Protocol (KMIP), an industry-standard communications protocol linking key-management systems and encryption systems. This protocol is governed by the Organization for the Advancement of Structured Information Standards (Oasis). Enterprises using multiple encryption solutions, each with their own respective (and often proprietary) key managers, have faced ever more complexity and cost. By implementing via an industry standard, the economics shift. Thus, KMIP is rapidly becoming the recognized method to manage encryption keys, no matter where they are required—and both vendors and their customers are committing to the technology.
KMIP isn’t entirely new. In fact, it was initially submitted to Oasis for standardization back in 2009. Enthusiastic vendors quickly began to announce updates to their products that would incorporate KMIP; actual demos and compliant-product deliveries followed.
With KMIP, a server stores and controls keys, certificates and user-defined objects. Client devices access these objects using the protocol through a server-implemented security model. Objects have core base-object properties such as key length and value, as well as extended attributes that can include user-defined features. Objects each have a unique, fixed identifier along with a name, which can be altered if desired.
Momentum is building. According to Oasis, the 2013 RSA Conference included KMIP clients from Cryptsoft, IBM, Quintessence Labs, and Thales e-Security communicating with key-management servers from Cryptsoft, HP, IBM, Quintessence Labs, Thales e-Security, Townsend Security and Vormetric. The clients and servers were able to demonstrate the full key-management life cycle, including creating, registering, locating, retrieving, deleting and transferring symmetric and asymmetric keys and certificates between vendor systems. Support for older and newer versions of KMIP was also demonstrated at the event.
Although KMIP has provided broad enablement for encryption through simplified key management, others have been focusing on delivering innovations related to the inherent performance challenge—particularly relative to backup activities, which can become unacceptably slow when coupled with encryption and decryption activities. These performance challenges can inhibit the ability to implement encryption as well as perform other vital tasks. For instance, traditional backup systems often lack sufficient scalability and robustness to simultaneously meet the challenges presented by explosive data growth, increasing risk and complexity, tight budgets, and the additional requirements of encryption—especially simultaneously. Data growth in particular—generally averaging about 20 percent per year—magnifies all the other challenges.
One part of the challenge is simply having enough storage. But facing a need to backup data from powerful applications such as enterprise resource planning (ERP), databases like SAP, and Oracle Business Suite, large enterprises and their IT staff are recognizing that just adding more non-scalable disk solutions or tape libraries to the mix is no longer a cost-effective or efficient strategy. They need approaches that are more workable; providing enterprise-optimized data-protection solutions that can help them address these challenges and stay ahead of the data growth curve.
Enterprises need a powerful and flexible data-protection platform that delivers scalability, performance and flexibility to accommodate both data growth and changing storage architectures, as well as enhanced data-protection features such as encryption. Furthermore, they must be reliable in addition to being easy to deploy and use, providing higher levels of service while reducing operating and capital costs.
Successful solutions to the performance dilemma, as outlined above, combined with encryption and the capabilities of KMIP, should deliver performance that neither suffers when data is being “ingested” nor increases the “time to safety” (the amount of time when data is on its way to disk but not yet encrypted).
Enterprise-scale organizations must move ahead now, gaining the benefits of more-capable and efficient storage infrastructure along with the crucial protection that encryption of data at rest provides. Doing so can help contain costs and avoid the alarming risks associated with data breaches and compliance violations. Fortunately, the technology is available today.
One of our customers, a very large natural gas company is a
Symantec NetBackup user with huge data volumes to protect. Like many Symantec
customers, they see the advantages of Symantec’s OST products and have started
a migration process from their current NBU deployment.
Change is constant in large enterprises and they appreciate the value of
systems and technologies that can accommodate change with minimal cost or
upheaval to ongoing operations. As this company moves its backup from a
NetBackup environment with Fiber Channel attached SAN storage to a
NetBackup/OST environment with 10 Gigabit Ethernet, the need for a scalable
disk based data protection platform with multiprotocol flexibility is acute.
Instead of adding multiple bounded appliances and dividing their backup volumes
between them, they are using a flexible, scalable appliance that uses storage
pooling technology to separate their NBU and OST environments. The platform
allows them to add compute and storage resources to their environments as
needed and to migrate from one protocol to the other in a seamless, risk-free
Flexibility and agility goes a long, long way in large enterprises.
The volume of data that every enterprise is dealing with continues
to grow at an alarming rate. Backup has been on the forefront of this wave. In
fact, the amount of data being protected is growing so fast and the solutions
are so broken it’s like pouring gasoline on a fire. But this is just the
beginning – as technologies like cloud and big data continue to take hold, the
problem is getting significantly worse – and most of the traditional storage approaches
will fail under the weight of this growth. The “big data” wave will ripple
through enterprise environments starting in primary storage and building to a
tsunami of epic proportions by the time it reaches the data protection
environment. It will create fire-drill after fire-drill.
Backup is the first place customers will feel the pain – any
non-scalable data protection technology will leave customers with a hard choice
– keep buying difficult to manage data protection silos or switch to truly scalable
data protection architectures that have a chance of solving the problem. “Big
Data Backup” is about to hit us all and break the infrastructure we’ve been
building for the last decade.
But the trend won’t stop there. Current methodologies for archive
and staging data for traditional analytics platforms will also fail as data
sets grow to sizes that turn simple data transfers into major pain points. The
traditional scale out NAS approach just won’t keep up. Newer massively scalable
object repositories will become the only viable solutions going forward. Any
second tier solutions not built around these contemporary technologies will
eventually fall out of competitive positions.
As we already are seeing, the pain around building scalable
primary cloud and analytic repositories is already being felt. Most enterprise
customers simply can’t approach the problem like Google, Yahoo and other
engineering dominant companies. Massively scalable and manageable analytic
platforms solutions are needed -- and they must interoperate efficiently with
similarly designed scalable data protection platforms going forward.
Interoperability is not just a nice-to-have, it is essential.
Really big data needs to move as little as possible, and when it does move,
efficiency will become critical. Most of the traditional analytics storage
technologies are buckling under the current load. With no slow-down in sight
we’re sure to see these traditional technology approaches fail completely in
the near future.
So what is a customer to do? First, be aware of the size and scope
of the problem. It’s overwhelming and it’s going to get much, much worse.
Second, be cautious about committing to tactical fixes and traditional siloed
approaches. Even if they let you “get by” for a while, they will eventually
fail on all fronts – causing massive added cost, weak manageability, and data
movement that will crush you.
Start to evaluate emerging “Big Data Backup”, “Big Archive” and
“Big Data” solutions and build plans to integrate them into their storage
ecosystems soon. And finally, customers need to remember emerging solutions
need to not only scale and perform cost effectively, but they also need to be
manageable in all ways. IT staff are being asked to manage more and more data
per person. In fact the ratio of IT staff to volume of data managed is getting
ridiculous, so issues like ease of deployment, upgrade and normal operation are
critical. Remember, these solutions must also automate data management and
integrity tasks, periodically scrubbing data. The bigger data stores get (and
they will get huge in coming years), the more data integrity on these systems
becomes an issue to worry about.
In a coming blog, I will discuss the demands that big data is
making on backup environments in more detail.
Big Data Demands Big Data Backup
As I mentioned in my last blog Warning: Big Data will Break Your Backup, we
have all gotten used to hearing about huge data growth and its impact on data
centers. As daunting as this data growth has been, it doesn’t compare to the
tsunami being generated in Big Data environments. Big data is measured in
petabytes and growing by orders of magnitude annually. Simply adding more
traditional data protection technologies – tape and multiple siloed systems is
not sufficient. They are just too slow, and too labor-intensive be feasible and
cost-effective in these environments.
Big Data Environments Need a New Approach
Just as big data primary storage is efficient, fast, and scalable, big data
secondary storage and data protection needs to be efficient, fast, and scalable
too. Let’s start with efficient.
With big data protection comes big costs. The days of backing up
and storing everything (just in case) are over. Cost reduction, labor savings,
and system efficiency are paramount to allow organizations to protect massively
growing amounts of data while meeting backup windows and business objectives.
Automatic, tiered data protection that includes CDP, snapshots, D2D, tape,
remote replication and cloud is a must-have. Data protection technologies
designed for big data environments need to be capable of automatically
identifying and moving low-priority data to the lowest cost recovery tier -
without administrator involvement.
Tiered Storage Becomes Essential
A tiered recovery model enables data managers to balance costs with recovery
time, depending on the recovery SLA requirements of the specific dataset. While
low priority data may have slower restore times, all restores use the same
mechanism, regardless of tier.
Second, even with smart policy driven tiering, big data
environments still need fast, scalable ingest performance to meet their data
protection needs. Stacking up dozens of single-node, inline deduplication
systems will lead to crushing data center sprawl and complexity. To meet data
protection needs and windows, big data environments need massive single system
capacity coupled with scalable, multi-node deduplication that doesn’t slow data
ingest or bog down restore times.
Compatibility with Existing Infrastructure and Reporting
Third, given the scale of big data environments, high risk rip and replace
solutions are out of the question. New data protection solutions need to
coexist and integrate seamlessly into existing environments without disruption
or added complexity. Optimally, IT staff can manage the resulting environment -
both new and existing IT resources - in one simple view that consolidates
management and reporting for the entire infrastructure.
Database Deduplication Efficiency
Big Data environments often have a large volume of duplicate data in segments
that are too small for many of today’s deduplication technologies to detect.
New big data-optimized deduplication technologies are required to minimize
costs and footprint regardless of where the data is retained.
The bottom line: traditional data protection cannot handle
demanding big data environments. A new approach is needed that delivers:
• Fast, deterministic ingest/outgress performance
• Intelligent, automated tiering
• Enterprise-wide management control and reporting
• Scalable deduplication designed specifically for big data database
In a coming blog, I will cover the Big Data topic in more depth
with a discussion of what is needed in a big backup appliance.
Large enterprises must look closely at the strengths and
weaknesses of the available deduplication technologies on the market before
choosing a solution that best meets the needs of their own Big Backup
The volume of data generated by most companies is now growing at
such an explosive rate that many data centers have simply run out of the space,
power, cooling and storage capacity required to handle it. In large enterprise
organizations with ‘Big Backup’ environments, the sheer volume and variety of
data to be protected requires a level of performance and scalability that few
data protection technologies can deliver.
Keeping up with exponential data growth within budget, space,
power and cooling constraints is a constant data center challenge. And it is
being compounded by increasingly stringent regulatory requirements and business
initiatives demanding higher service levels, longer online retention times and
higher levels of data protection. There are several deduplication technologies
available to meet the needs of small-to-medium-sized organizations. However, for
large enterprise organizations with Big Backup environments, most deduplication
technologies fall short, resulting in costly, inefficient capacity reduction,
and time-consuming administration. Understanding the distinctions between these
technologies is essential for choosing those most appropriate for Big Backup.
SEPATON recommends that all large enterprises ask their backup
technology vendor the following ten questions. This will help them gain a clear
understanding of the strengths and drawbacks of each option so that they can
choose one that best meets their Big Backup requirements:
1. What impact will deduplication have on backup performance, both
now and over time?
2. Will deduplication degrade restore performance?
3. How will capacity and performance scale as the environment grows?
4. How efficient is the deduplication technology for large databases (e.g.
Oracle, SAP, SQL Server)?
5. How efficient is the deduplication technology in progressive incremental
backup environments such as Tivoli Storage Manager (TSM) and in NetBackup OST?
6. What are realistic expectations for capacity reduction given the high data
change rate common in Big Backup environments?
7. Can administrators monitor backup, deduplication, replication and restore
8. Can deduplication help reduce replication bandwidth requirements for large
enterprise data volumes without slowing backup performance?
9. Can IT teams ‘tune’ deduplication by data type to meet their specific needs?
10. How much experience does the vendor have with large enterprise backup
applications such as Symantec NetBackup/OST and TSM?
For more details, a white paper titled 'Choosing an Enterprise-Class Deduplication Technology.'
Virtual Strategy Magazine invited me to reflect on the past year in a recent articlethat looks ahead to the data protection challenges, trends, and emerging technologies that will shape our coming year. As IT professionals, we divide our time between looking back -- analyzing our systems to identify ways to improve them, to save money, to do more with fewer resources – and looking ahead to the latest technologies to keep our data centers efficient, flexible, and responsive in today’s dynamic marketplace. Here are trends to watch out for in 2014:
Trend #1: Data Growth Hockey Stick
Data growth is nothing new in enterprise data centers. What will be new for 2013 is the hockey stick growth that will be driven by increased use of big data analytics, growing number of applications, and increased use of large database-driven applications.
What this means:
Rapid data growth causes three major (and myriad minor) issues. First, backing up large and fast-growing data volumes within backup windows becomes an almost daily struggle.
Second, your costs skyrocket. Capital spending increases as you quickly (and often unexpectedly) outgrow your backup system and start adding systems for capacity and performance. Your IT labor costs go up too as your IT administrators spend hours moving data off disk-based systems onto tape archives, adding and load balancing additional systems, and tuning every possible part of your data center to squeeze out every last drop of performance.
Third, you start compromising protection levels. You can’t move massive volumes of data offsite for disaster protection efficiently, you can’t backup data as frequently as you would like. You can’t encrypt data at rest.
What you should do:
Use systems that are specifically to handle enterprise data protection needs. These systems lets you add performance and capacity as you need it to protect massive data volumes without data center sprawl. They automatically load balance and tune themselves for optimal efficiency. They are built to deliver the performance you need to meet backup windows with ease.
Trend #2: Increased Reliance on Large Database-Driven Applications in a Sea of Unstructured Data
Enterprises are moving more and more business operations onto large Oracle, SQL, DB2 database driven systems. As a result, enterprises are backing a wide mixture of data types and more database data than ever before and their tolerance for downtime for this data is nearly zero.
What this means:
Database data is notoriously difficult to deduplicate efficiently for two reasons. First, databases store data in very small segments (<8KB) that inline deduplication systems cannot process without slowing backup performance. Second, to meet backup windows, they are typically backed up using multiplexing, and multistreaming, which are not supported by inline, hash-based deduplication systems. As a result, data centers are being inundated with massive volumes of under-deduplicated database data.
What you should do:
Use a data protection system that is built to handle enterprise class databases and backups with a mixture of data types. These systems use smart hybrid deduplication technology that determines the best technology to use based on the actual data type. They can deduplicate any data type, including database data – even in multiplexed, multistreamed backups - at the byte level without slowing ingest performance. They analyze the composition of incoming data streams and apply the deduplication technique (inline and/or post process) that will deliver the most efficient combination of performance to meet backup windows, capacity optimization, and replication bandwidth use.
Trend #3: Increasingly Complex Backup Environments
The days of simply backing up to tape or virtual tape are long over. Today’s data protection environment uses a wide range of backup protocols and data protection technologies– from backing up to cloud, tiering of data, snapshots, source and target deduplication, archiving, to name a few.
What this means:
Data center complexity quickly translates to increased cost and more risk of data loss and downtime. As data centers continue to divide backup data volumes into disparate backup and protection schema, they not only lose efficiencies of scale, they also increase the likelihood of human error. As we look ahead to the coming months and years, data centers will continue to grow and increase in complexity as they introduce more different technologies and backup schemes to handle data growth and meeting regulatory requirements and service level agreements.
What you should do:
Consider implementing a data protection system that is designed to provide the best efficiency and most flexibility that you require today and also into the future. These systems can read the content of incoming backup volumes, make intelligent decisions in regards to best technology for efficiency, and move massive data volumes with remarkable efficiency. These systems are designed with the future in mind for enterprise data centers as they will provide a way for IT managers to leverage the benefits of leveraging numerous data protection technologies – cloud, tiering, hybrid deduplication, snapshots, etc. - without adding risk and complexity.
These three trends will shape the future of enterprise data protection in 2014 and beyond.
Data growth continues to skyrocket and data is becoming increasingly important to business. As a result, the data center infrastructure to support that growth is becoming increasingly important. Yet with constrained IT budgets, backing up increasing data volumes is even more difficult and managing the increasing diversity of applications, systems, protocols, and types of data sources is even more complex.
One of our customers, NTT Europe
, a managed service provider, was planning to expand throughout Europe. They needed a backup environment that could grow with them without disrupting service to their end users. They chose Sepaton because of our grid scalability and because of our unique deduplication capability that uses recently patented technology to deduplicate very large data volumes efficiently without slowing backup or restore performance.
According to NTT Europe, other deduplication methods were much slower and less efficient in reducing capacity requirements than the Sepaton DeltaStor™ technology. Using our patented deduplication technology, NTT can now restore customer data immediately, without waiting for lengthy ‘rehydration’ processes. Check out their story here
While deduplication ratios vary based on data types, it’s fair to say our customers have seen dramatic space savings using Sepaton deduplication methodologies and are saving costs by backing up entirely to one easy-to-manage data protection appliance. Many of our customers have seen industry-leading 10:1 data reduction in their progressive incremental TSM backup environments.
For example, with Hitachi Data Systems (HDS)
high density drives in the Sepaton system combined with DeltaStor deduplication, one of the country's largest regional healthcare systems has enough capacity to support their data growth for approximately three more years. They anticipate that the added density will save them thousands of dollars annually and now have the capacity to pull all data off their existing physical LTO4 drives and backup entirely to one easy-to-manage Sepaton data protection appliance.
As the volume of data continues to grow, new, more cost effective ways of protecting data are needed and innovative and intelligent deduplication is an important way to go about doing just that.
I recently provided content to eWeek for a slideshow on 10 best practices for implementing backup and recovery.
Data protection capabilities for high-performance, virtualized environments come with some of the thorniest issues in IT. Problems exist in how well backup and recovery practices work in both physical and virtual realms—because most enterprises now have both—as well as in cost and complexity. In a recent study published in eWEEK, which surveyed businesses in the United Kingdom, the United States, Germany and France, 85 percent of small and midsize businesses (SMBs) cited cost as an obstacle, 83 percent are concerned about their data protection capabilities, and 80 percent pointed to complexity issues. These challenges are based partly on continued attempts to apply traditional, physical-world strategies to data protection in virtual environments, instead of adopting a new approach. Enterprises are taking new approaches to the challenges they face in implementing data protection. For example, 55 percent of SMBs say they plan to change their tools for backing up virtual environments in the next two years. Here are 10 key recommendations for implementing a high-performance, cost-effective, grid-scalable backup and recovery system with high deduplication rates.
Data Center Knowledge published a recent article of mine on the hidden costs of system sprawl.
While data center managers have grown accustomed to rapid data growth, few could have anticipated the unprecedented data growth and increased complexity that has overwhelmed many data center backup environments in the past few years. According to industry analysts, data in large enterprises is growing at 40-60 percent compounded annually.
Data growth is fueled by the proliferation of new business applications, the introduction of Big Data analytics, the increased use of mobile devices and tablets in the work place, and the increased use of large databases to run core company functions (ERP, payroll, HR, production management). Companies are not only creating massive volumes of data, they are also under pressure to meet increasingly stringent and complex requirements for protecting and managing that data. For example, they have to back it up in shorter times, retain it for longer periods of time, encrypt it without slowing backup performance, replicate it efficiently, and restore it quickly.
Until recently, many enterprise data managers responded to data growth by simply adding disk-based backup targets. The most common type of disk-based backup target provided inline data de-duplication and a reasonable level of performance and capacity to accommodate the increased data volume. However, these systems are simply not designed for today’s massive data volumes or fast data growth because they lack two critical capabilities: they do not scale and they do not de-duplicate enterprise backup data efficiently. As a result, for many large enterprise data centers, the “add another system” approach has reached its breaking point.
The Hidden Costs of Sprawl – Total Cost of Ownership
For many organizations the breaking point for non-scalable systems is the point at which they cannot any longer meet their backup windows. While adding a single system may not seem overly cumbersome, for large enterprise data centers that require several of these systems, it can add unplanned cost, complexity, risk, and administrative time. The hidden costs and total cost of ownership (TCO) impact are significant:
Overbuying systems. Companies are forced to add an entire system when they have plenty of capacity but only need more performance or conversely, have performance and need more capacity.
Wasting money on capacity. By separating data onto multiple non-scalable systems, these systems cannot de-duplicate globally, reducing the efficiency of their capacity optimization.
Wasted IT admin time. To add a new non-scalable backup system, IT admins have to divide the existing backup(s) onto multiple new systems and load balance for optimal utilization a process that becomes more time-consuming and complex with every new system added.
Added maintenance cost. Each new system increases the cost of system maintenance by an order of magnitude. Every time a new software or hardware update or upgrade is needed, or standard maintenance is required.
Slow backups. Non-scalable systems typically use hash-based, inline de-duplication that slows backup performance over time. They are highly inefficient in database backup environment common in enterprises data centers for two reasons. First, databases often store data in sub-8KB segments that are too small for inline, hash-based deduplication to process efficiently without becoming a bottleneck to backup. Second, they do not support fast multiplexed, multi-streamed database backups – requiring IT staff to choose between fast backups and capacity optimization.
Rising data center costs. In simple terms, more systems with less-efficient de-duplication means more rack space, power, cooling, and data center floor space.
Less is More for Low TCO
In today’s fast-growing enterprise backup environments, consolidating backups onto a single, enterprise-class disk-based backup appliance is proving to be both more cost-efficient and less prone to human error and data loss than the “siloed” approach described above.
Backup and recovery appliances are designed specifically to handle the massive data volumes and complex backup requirements of today’s data centers. These purpose built backup appliances (PBBAs) are designed to backup, de-duplicate, replicate, encrypt, and restore large data volumes quickly and cost efficiently. To ensure you choose an enterprise-class backup and recovery appliance, ensure to use the following best practices:
Opt for Guaranteed High Performance
Understand the performance impact that processing-intensive functions, such as de-duplication, replication, and encryption have on the system. Enterprise-class systems are designed to perform these functions in a way that does not slow performance. Some even offload CPU functions Ensure that any published performance rates are for guaranteed, continuous performance, and not simply the highest rates achievable in a widely varying ingest rate.
Grid Scalability is Essential
As described above, adding, managing, and using multiple backup systems is not practical or cost-efficient in today’s fast-growing, complex data centers. Enterprise-class backup and recovery systems offer grid scalability, that is, the ability to add performance and/or capacity independently as you need it. This pay-as-you-grow model eliminates over-buying, reduces IT management time, and enables you to store tens of petabytes of data in a single, consolidated backup appliance.
Storing data in a single, optimized system has the additional benefits of enabling highly efficient, global de-duplication, and eliminating the need for load balancing and ongoing system-tuning.
Ensure Deduplication is Designed for Enterprise Data Centers
One of the most effective ways to reduce the cost of backup and recovery is to implement enterprise-class de-duplication. Unlike de-duplication optimized for small-to-medium-business, enterprise de-duplication is designed to enable faster backup performance and better overall capacity requirements. It is also capable of tuning de-duplication to the specific data types for optimal use of CPU, disk, and replication resources. For example they can de-duplicate database data at the byte level for optimal capacity savings or recognize data that will not de-duplicate efficiently (i.e., image data) and back it up without de-duplication. This “tunability” can save enterprises thousands of dollars in savings in capacity and processing costs.
Reporting and Dashboards Enable Savings
Detailed reporting and dashboards are key to enabling IT administrators to manage more data per person. They automate all disk subsystem management processes and put detailed status information at the administrator’s fingertips. They also provide predictive warning of potential issues enabling administrators to take action before they become urgent.
Lowest Total Cost of Ownership
For today’s large enterprise backup and recovery environments, the days of adding more and more backup systems are over. The speed of data growth, massive volume of data, and complexity of backup and recovery policies necessitate the use of enterprise-class purpose-built backup appliances. These appliances enable organizations to maintain backup windows by moving massive data volumes to the safety of the backup environment at predictable, fast ingest rates. They also streamline and simplify complexity by consolidating tens of petabytes of stored data onto a single, cost-efficient easy-to-manage system.
Virtual Strategy Magazine recently featured an article I wrote entitles "Enterprise Data Protection 2013."
I write that for the past ten years or more, large enterprises have struggled to keep up with exponential data growth. This growth has driven enterprises to move from physical tape to disk-based virtual tape libraries, and now to more powerful disk-based data protection solutions than ever before. Despite these improvements data growth continues to push backup and recovery systems to their limits.
In 2013, I predict that the exponential data growth that we have experienced in the last ten years will explode to a level that takes many IT departments by surprise. Driven by Big Data analytics, cloud, and increased reliance on very large databases (Oracle, DB2, SAP), data growth is about to increase at rates that most data centers cannot accommodate.
Data protection in this fast-growing environment will quickly push backup systems with inline, hash-based deduplication to the breaking point. Without the ability to scale, enterprise data centers will find themselves first adding more and more of single-node backup systems to get the performance they need to meet backup windows. As these systems multiply like rabbits in the data center, administration labor costs will begin to rise. More systems mean more system maintenance, updating, and tuning – and more chance for human error.
Evolution to Data Protection Consolidation on Grid Scalable Systems
Enterprises will increasingly move to consolidate these systems onto one or a small number of scalable backup and recovery appliances that not only scale performance and capacity but also provide a holistic view of the backup environment companywide.
To meet data protection needs and shrinking backup windows, enterprises will move away from backup and recovery solutions that use pure hash-based inline deduplication and to more capable Content-Aware multi-method deduplication technologies that can scale both performance and capacity as the backup environment grows. They also designed to cut capacity in multiplexed, multistreamed databases and massive backup environments without slowing backup performance. Companies will be looking to adopt systems that provide a high degree of automation of disk subsystem management (e.g., load balancing), and provide powerful management and reporting interfaces that enable IT admins to track data and manage data through the backup, deduplication, replication, and secure erasure processes efficiently.
Companies will adopt new data protection solutions that coexist and integrate seamlessly into existing environments added complexity. They will enable companies to manage all of their data protection through a simple management and reporting for the entire infrastructure – including both existing and new technology. Consolidation also enables more efficient deduplication and overall capacity optimization as data is not divided among numerous “silos of storage”.
Although today’s enterprises are accustomed to rapid data growth, few are prepared for the enormous growth of data that is going to result from big data analytics and very large databases. They will move away from the single node systems and toward more sophisticated, grid scalable technology designed to backup, deduplicate, replicate, and restore tens of petabytes of data on a single system.