1. read speed, 2. write speed, 3. throughput, 4. data protection/integrity, 5. de-dupe topology - at target, at source, both - for last two check the impact on the source 6. the need for agent/plugin installation - check FW requirements (ports to open etc.) 7. space reclamation (garbage collection, filesystem cleaning etc.) - check if the system will be able to finish GC before next run. 8. ability to scale the system performance and capacity 9. data transfer protocols (tcp/IP, FC, iSCSI etc.) 10. application/backup software interoperability - for source based de-dupe, for additional services like virtual synthetic full backups
Find out what your peers are saying about Dell Technologies, Hewlett Packard Enterprise, NetApp and others in Deduplication Software. Updated: March 2024.
What is deduplication in networking? Deduplication is the process of eliminating duplicate copies of data from a system. Data deduplication improves storage utilization and can be administered in both data backup and network data schemes. Often called single-instance storage or intelligent compression, data deduplication optimizes your data backup storage by ensuring that only one instance of data is copied and stored.
I think the most important features to look for are whether dedupe is online or at rest and if the block size is fixed or not.
You should look at these feature and parameters:
1. read speed, 2. write speed, 3. throughput, 4. data protection/integrity, 5. de-dupe topology - at target, at source, both - for last two check the impact on the source 6. the need for agent/plugin installation - check FW requirements (ports to open etc.) 7. space reclamation (garbage collection, filesystem cleaning etc.) - check if the system will be able to finish GC before next run. 8. ability to scale the system performance and capacity 9. data transfer protocols (tcp/IP, FC, iSCSI etc.) 10. application/backup software interoperability - for source based de-dupe, for additional services like virtual synthetic full backups
recovery performance, data availability and accessibility
Performance penalty, on data read operation.
Impact upon performance. Is deduplication going to disrupt the workflow.
data availability , accessibility and performance.