Data deduplication is the process of preventing duplicate data from being stored and archived. This reduces the rate of data growth in an organization thereby significantly easing its management and lowering overall TCO.
In this guide:
- Data deduplication basics
- Business benefits of data deduplication
- Data Deduplication disadvantages
- Key product evaluation considerations
- Data deduplication vendors and their offerings
Duplicate data results in increased storage, backup and retrieval time and costs. Multiple copies of data in secondary storage systems are the biggest source of this type of data. The greater the amount of data in an organization, the more difficult it is to meet RTO (recovery time objectives) and RPO (recovery point objectives). Data deduplication helps address this by preventing duplicate data from being part of the equation.
Backup applications benefit from deduplication as most of the data in a given backup doesn’t change from the previous backup of an existing file system. Virtual environments also benefit quite a bit as marginally separate system files of different virtual machines can easily be combined into a single storage space.
Effectiveness of deduplication solutions is measured in terms of the de-dupe ratio as follows:
De-dupe ratio = Total data sent by the application(s) to the storage system/Total ‘raw’ storage capacity of the storage system
Deduplication solutions have been applied mainly to secondary storage systems in the past. Vendors are now augmenting their existing solutions or offering new ones to apply deduplication to primary storage systems as well.
1) Substantial reduction in data center capital and operating costs by lowering the requirements for equipment, power, cooling and floor space.
2) Recently, tapes have been replaced by disks as secondary storage media. Deduplication slows the need to purchase new disks.
3) Enabling more data to be backed up.
4) Increasing disk efficiency, thereby enabling longer disk retention periods.
5) Helping optimize use of network bandwidth by reducing the data sent on the wire for backup, replication and DR.
1) Performance impact – Data deduplication employing a fixed-block approach on primary storage may adversely impact performance. This is because of the inherently random nature of primary data leading to multiple chunks written at different locations, thus increasing data re-assembly times.
2) Loss of data integrity - Block-level deduplication solutions utilizing hashes create the possibility of hash collisions (identical hashes for different data blocks). This can cause loss of data integrity due to false positives, in the absence of additional in-built verification.
3) Backup appliance issues – Data deduplication requires a separate hardware device, often referred to as a “backup appliance”. This has a cost, and can lead to performance degradation if data is required to be transferred from backup server to the hardware device.
4) Performance impact due to shared resources - If a deduplication solution uses the hashing technique and shares computational resources on a device provisioning other services, overall performance is affected.
Do you really require a deduplication solution? This decision would be based on internal diligence about type and nature of data redundancy, data change frequency, full backup frequency, RTO and RPO objectives. Once you have determined that a data deduplication solution is required, consider the following aspects:
1) What levels of de-dupe ratios does the deduplication solution vendor claim? These claims must take into account the data set size as well as the point at which data deduplication is initiated, with linked assumptions about available processing power and network bandwidth.
2) Which of the three deduplication methods would suit your organization’s requirements?
a) Block-level - This method operates by creating treating data as “blocks/chunks” of data. Only one unique instance of the data is retained on secondary storage media. Hashes for each block are computed and stored in a hash lookup table. When new data enters the system, block hashes are computed and compared with entries in the hash lookup table. In case of a match, block data is replaced with reference to the existing block. An entry is made in the hash table for all other cases. Smaller chunks increase the probability of finding redundant data.
b) File-level - Works best for data that doesn’t change often, such as employee profile. Each duplicate instance is referred back to the single, unique file copy.
c) Content-aware deduplication - Content is semantics of the data being duplicated. This helps determine resources (processor/memory/IO) required for data deduplication.
3) Inline or post-storage? - To reduce storage requirementsinline mde may be used, wherein the deduplication method is applied on incoming data. This mode is better suited for block-level methods. Alternatively, in the post-storage mode, incoming data is stored first and analyzed later for duplicates. Prefer this method in case the deduplication appliance also doubles up as an NFS or CIFS share. This will ensure zero impact on application response times or access of shares by users.
4) Conduct systematic capacity planning of storage required for data recovery. This is essential when considering recovery of de-duplicated data created from primary data snapshots, as the deduplication process expands the data to its original size and not the snapshot size.
5) Determine the type of device your existing backup application expects the deduplication solution to be: VTL (Virtual Tape Library), tape or disk.
6) Does the deduplication solution support both NAS and SAN server connections?
7) Does the deduplication solution support deduplication of data from any type of application?
8) Does the solution move de-duplicated data or duplicate data for replication across the wire?
|Deduplication tool vendor||
Solution name and features
IBM ProtecTIER - Inline deduplication method ensures backup windows are adhered to without leading to any disruptions. Use of non-hash-based approach preserves data integrity.
Data Domain DD140 - Used for data deduplication in remote locations with replication for meeting RTOs.
CA ARCserve Backup - Employs target-based deduplication method to compare new backups with stored data at the block-level.
NetBackup 7 - Offers single console for multi-site monitoring, analytics and reporting.
Data ONTAP - Deduplication and compression are core components of NetApp’s operating architecture, implementable across multiple applications and storage systems.
|HP||HPStoreOnce - Deduplication technology used speeds up backups up to 4TB/hr and cuts storage capacity requirements by 95%. Can be used to centrally manage data center and remote office backups.|
* Further reading
Tutorial: Data deduplication tutorial