Atlantis: Putting a leash on storage expansion
Understanding the real impact of deduplication is critical for any IT professional in the modern datacenter. This is true from both performance and capacity perspectives. Choosing the right storage solution is now a longer-term commitment. Choose wisely by understanding the impact of a key data service functionality such as deduplication.
In the past deduplication, or dedupe, was a feature highly priced and only available from a small number of vendors. However, as the IT community began to see the explosion of data created and used by their machines, business consumers and applications, dedupe went from a nice to have feature to a critical data service. It was critical not only for data management but also for storage cost mitigation. These days, however, the feature is provided by both hardware and software vendors alike. The vendors each have their own methods and levels of success. Making it seem that deduplication is just a generic check box from a feature perspective, which applies to all modern storage solutions. However, when you really look under the covers this isn’t true at all. There is a lot fluff, FUD and downplay around deduplication in the industry, especially from those vendors that don’t have a strong story around this feature set. In this blog post I will articulate why de-duplication is important for your modern datacenter and give a deduplication technology deep dive.
Currently, deduplication effectiveness boils down to three variables:
What amount of data is analyzed at a specific time
When that analysis is done
Where it is done in the IO stream
Read the entire article here, Putting a leash on storage expansion