Inside TDV - The Data Vault Blog
Why ‘Big Data’ Deduplication Should be a Priority
The term “Big Data” is being thrown about more than ever these days. Is anyone else curious about just how much of data sprawl is comprised of duplicate data?
Someone was. A new study states that companies worldwide are now spending $44 billion on storage hardware and software to store and manage totally unnecessary copies of enterprise data.
That seems a bit unbelievable, but the study also claims that 85 percent of storage hardware purchases and 65 percent of storage software purchases are made just to support this duplicate data. (The research was based on market revenue and forecast data, along with a survey of more than 700 IT managers worldwide.)
Big data is challenge enough, but big redundant data?
How much duplicate data is your system dragging? And how much time and money is it costing your company? With TDV Cloud, all of our customers begin with a detailed analysis of their network to determine the amount of data that will be migrated to our cloud-based solution.
That process involves a deduplication of the data, which ensures that only vital files are retained. The software looks for the same data queued for backup more than once, regardless of if the files are on different servers or have different names. Common data is then backed up to the appropriate repository with a pointer from the files’s original location.
Our backup life cycle management then allows you to move old backup data to lower cost storage, thus aligning the value of your data with the cost of protecting it. TDV Cloud can not only can manage duplication but also ensure minimum capital outlay and an all inclusive monthly service charge.
This can save your company a ton of money and manpower, especially if your version of big data includes redundant data, or data that may be better served being archived or backed up only monthly.
A great example of this way to manage big data can be found over at ComputerWorld.com in Peter Eicher’s excellent blog post titled, “If files were bricks, you’d change your backup strategy.”
Literally, if you are moving files around daily in data tapes or via bare-metal backups, you’re essentially moving big data that doesn’t need to be constantly moving. What if you could select the files that need to be backed up daily or hourly, or even more frequently for truly critical data?
That kind of solution makes big data seem like a much lighter burden.