Data deduplication
Terms related to data deduplication, including definitions about single-instance storage and words and phrases about eliminating redundant data.DAT - WHA
- data archiving - Data archiving migrates infrequently used data to low-cost, high-capacity archive storage for long-term retention.
- data at rest - Data at rest is a term that is sometimes used to refer to all data in computer storage while excluding data that is traversing a network or temporarily residing in computer memory to be read or updated.
- data center - A data center -- also known as a datacenter or data centre -- is a facility composed of networked computers, storage systems and computing infrastructure that businesses and other organizations use to organize, process, store and disseminate large amounts of data.
- data deduplication - Data deduplication -- often called intelligent compression or single-instance storage -- is a process that eliminates redundant copies of data and reduces storage overhead.
- data deduplication hardware - Data deduplication hardware is disk storage that eliminates redundant copies of data and retains one instance to be stored.
- data deduplication ratio - To calculate the deduplication ratio, divide the capacity of backed up data before duplicates are removed by the actual capacity used once the backup is complete.
- data federation software - Data federation software is programming that provides an organization with the ability to collect data from disparate sources and aggregate it in a virtual database where it can be used for business intelligence (BI) or other analysis.
- data lifecycle management (DLM) - Data lifecycle management (DLM) is a policy-based approach to managing the flow of an information system's data throughout its lifecycle: from creation and initial storage to when it becomes obsolete and is deleted.
- data reduction - Data reduction is the process of reducing the amount of capacity required to store data.
- data reduction in primary storage (DRIPS) - Data reduction in primary storage (DRIPS) is the application of capacity optimization techniques for data that is in active use.
- data scrubbing (data cleansing) - Data scrubbing, also referred to as data cleansing, is the process of amending or removing data in a database that is incorrect, incomplete, improperly formatted or duplicated.
- global deduplication - Global deduplication is a method of preventing redundant data when backing up data to multiple deduplication devices.
- inline deduplication - Inline deduplication is the removal of redundancies from data before or as it is being written to a backup device.
- KSM (kernel samepage merging) - KSM (kernel samepage merging) is a Linux kernel feature that allows the KVM hypervisor to share identical memory pages among different process or virtual machines on the same server.
- memory mirroring - Memory mirroring is the division of memory on a server into two channels.
- post-processing deduplication (PPD) - Post-processing deduplication (PPD), also known as asynchronous de-duplication, is the analysis and removal of redundant data after a backup is complete and data has been written to storage.
- source deduplication - Source deduplication is the removal of redundancies from data before transmission to the backup target.
- target deduplication - Target deduplication is the removal of redundancies from a backup transmission as it passes through an appliance sitting between the source and the backup target.
- What is data management and why is it important? - Data management is the process of ingesting, storing, organizing and maintaining the data created and collected by an organization, as explained in this in-depth look at the process.