Data Efficiency for 21st-Century Data Centers

Data Efficiency for 21st-Century Data Centers

Storage efficiency has typically referred to the processes resulting in reduced storage and bandwidth capacity requirements. Compression, thin provisioning, data deduplication and even storage virtualization have had a huge impact on storage, IT efficiency and, ultimately, the total cost of ownership (TCO) of enterprise storage. These technologies are pervasive in data center services, such as production storage, backup, WAN optimization, and archiving.

In today’s post-virtualization data center, virtualized workloads with different IO streams are sharing the physical resources of the host. This results in random IO streams competing for resources—and the emergence of new efficiency requirements as the IOPS required to service virtual workloads has increased. Some band-aids applied to the IOPS problem include overprovisioning HDDs or investment in SSDs/flash—with both contributing to a higher cost per gigabyte of storage allocated to each virtual machine (VM).

Enter data efficiency. Data efficiency capitalizes on the familiar storage efficiency technologies, such as compression and deduplication, but executes them in a way that positively impacts both capacity AND IOPS—and costs—in today’s modern data centers.

To help you understand data efficiency better, our blog is designed to detail how data centers have evolved in the 21st century, what problems need to be addressed today, how using “yesterday’s technology” or “yesterday’s approaches” cannot solve the problem, and why it’s worthwhile to solve the problem.

Upcoming Topics to Include:

Data Efficiency Overview

Defining The Data Problem

CompressionDeduplication

Deduping IO

Hardware-Assisted Deduplication

Advantages of Data Efficiency: WAN Benefits

Advantages of Data Efficiency: Cache Benefits

Advantages of Data Efficiency: Data Protection Benefits

Data Efficiency Across the Lifecycle

Last Words on Data Efficiency