Massive volumes of data are being produced and transmitted every minute — data that needs to be stored, protected and quickly accessed. Video and other multimedia files are straining storage resources. Long-term data retention requirements necessitate greater storage capacity. Cloud architectures and highly virtualized environments must be backed up effectively. At the same time, backup windows are shrinking.
Of all the challenges organizations face when managing the storage environment, capacity is the primary concern. Data storage requirements never shrink, they only get exponentially bigger. Organizations need ways to keep a lid on capacity requirements while ensuring that users and applications have the storage resources they need.
That’s why data reduction strategies such as de-duplication are playing an important role in today’s data center environment. De-duplication is a technique that reduces the amount of data being stored by detecting repetitive patterns and eliminating duplicate data. Copies of files are replaced with a pointer to the original, shared file. For example, if a PowerPoint presentation is emailed to 50 people, de-duplication ensures that only the original presentation is stored, not 50 copies of the same file.
De-duplication has long been a critical part of the storage infrastructure, primarily for backup and archival. The resulting savings in storage costs and capacity have always been and continue to be significant. Depending upon the data being de-duped, savings as high as 95 percent can be achieved as capital, operational, and power and cooling costs are reduced.
At the same time, recent data center trends are causing organizations to rethink their de-duplication strategies. More organizations are using de-duplication for primary storage, particularly as they move to flash storage platforms. Tier 1 de-duplication must be carefully implemented and integrated to ensure performance and data integrity.
Big data also complicates de-duplication. High volumes of unstructured data and associated metadata that is suitable for de-duplication must be separated from data that is not. Historical data must be prioritized to enable agile storage and retrieval, which is quickly becoming a business necessity.
What should you look for in a de-duplication solution for the next-generation data center?
- Purpose-built de-duplication appliances. This technology should be easy to install, deliver high performance, and work with backup applications.
- Variable-block de-duplication. Although this approach is resource-intensive, it optimizes disk storage capacity as data grows and maximizes disk storage density.
- Pay-as-you-grow scalability. This approach will help prevent overprovisioning of storage capacity.
- Advanced monitoring and reporting capabilities. Effective de-duplication requires effective management tools and performance analysis.
- Security that doesn’t compromise performance. Encrypt data at rest and in motion to address vulnerabilities before a breach occurs – without causing performance issues.
De-duplication is growing in both complexity and importance as organizations seek to get a handle on today’s storage demands and better manage workflows that control data movement. Let Technologent help you develop and implement an effective de-duplication strategy that optimizes your storage environment and meets the demands of the modern data center.
Tags:
Data StorageOctober 27, 2015
Comments