Sorted deduplication: How to process thousands of backup streams

@article{Kaiser2016SortedDH,
  title={Sorted deduplication: How to process thousands of backup streams},
  author={J{\"u}rgen Kaiser and Tim S{\"u}{\ss} and Lars Nagel and Andr{\'e} Brinkmann},
  journal={2016 32nd Symposium on Mass Storage Systems and Technologies (MSST)},
  year={2016},
  pages={1-14}
}
The requirements of deduplication systems have changed in the last years. Early deduplication systems had to process dozens to hundreds of backup streams at the same time while today they are able to process hundreds to thousands of them. Traditional approaches rely on stream-locality, which supports parallelism, but which easily leads to many non-contiguous disk accesses, as each stream competes with all other streams for the available resources. This paper presents a new exact deduplication… CONTINUE READING

Similar Papers

Citations

Publications citing this paper.
SHOWING 1-3 OF 3 CITATIONS

References

Publications referenced by this paper.
SHOWING 1-10 OF 15 REFERENCES

Extreme Binning: Scalable, parallel deduplication for chunk-based file backup

  • 2009 IEEE International Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication Systems
  • 2009