Sunday, May 17, 2009

Data Domain claims to have the fastest global data de-duplication for enterprise DR readiness.

Earlier this month Data Domain announced some interesting enhancements to their Enterprise Replicator Software. The announcement underlined their market performance to justify their bullish exclamation that they are #1 in data de-duplication.

Note: I have updated this posting since its original publishing.

Originally introduced to the market in 2004:

  • By 2006 Data Domain had 330 customers that has grown to over 3000 today.

  • In 2006 they had shipped over 1000 systems, today this number is now over 8000.

  • In 2006 they had over 30PB of data under protection, today the number claimed is greater than 1000PB.

  • In 2006 the replication attach rate was 40%, now it is close to 60%.

The key elements of the announcement were:

  • Collection Replication. Basically a full system mirror between two systems for DR purposes.
  • Global cross-site de-duplication. Simply put a feature that collects de-duplicated data from multiple remote sites, all managed by a single directory of common objects that is representative of all these disparate data sources. This limits the replication to only unique data objects when viewed across all data sources. The benefit of this aggregated view is to significantly minimize the number of unique data objects replicated, hence minimizing expensive bandwidth requirements. The logical throughput performance quoted was 6GB/sec or 21.6 TB/hr.

  • The fan-in ratio has been increased to 90 to 1. Meaning that one central instance of a single Data Domain controller (DD660 or DD690) can now support the de-duplication of up to 90 remote data sources, significantly simplifying back-up and recovery actions.

Data reduction techniques such as data deduplication and data compression are now well accepted as legitimate techniques in the battle to control today’s explosive data storage growth. These logical representations of data have a significantly smaller physical footprint requiring not only less physical storage but significantly less bandwidth when replicated. These basic data reduction methodologies continue to evolve and there is a growing argument that the basic technologies will become commoditized within the next 12 to 18 months with those vendors who will be successful will be the thoughtful innovators who create meaningful, value added extensions to the basic data reduction technology.

Data Domain are certainly showing such innovative tendencies.