Improving the Performance of Primary Storage Systems Using Leveraging Data Deduplication
Data deduplication is a specialised data compression technique used in computers to get rid of redundant copies of repeated data. Intelligent (data) compression and single-instance (data) storage are words that are similar and partly interchangeable. This method may be used to increase storage efficiency and the quantity of data that has to be sent should be reduced over networks. Unique data chunks, or byte patterns, are found and saved throughout the deduplication process through a process of analysis. Recent studies have demonstrated that main storage systems in the Cloud offer moderate to high levels of data redundancy. It demonstrates that as a result of the small I/O requests to redundant data's comparably high temporal access locality, data redundancy displays a substantially greater degree of intensity on the I/O path than that on discs. The system in this project suggests using a performance-oriented I/O deduplication technique known as POD. It is used to increase the I/O performance of main storage systems in the Cloud without compromising the latter's capacity savings.