Data depublication using : Hashing algorithm / Naimah Nayan

Data depublication is method that help reduce the redundant data in storage capacity. With the rapid growth of digital data that generated in the digital world, the capacity of storage usage will increase rapidly. To achieve deduplication efficiency in system storage, the duplicate data need to be e...

Full description

Saved in:
Bibliographic Details
Main Author: Nayan, Naimah
Format: Thesis
Language:en
Published: 2019
Subjects:
Online Access:https://ir.uitm.edu.my/id/eprint/26848/1/TD_NAIMAH%20NAYAN%20CS%20R%2019_5.pdf
https://ir.uitm.edu.my/id/eprint/26848/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1833062647837229056
author Nayan, Naimah
author_facet Nayan, Naimah
author_sort Nayan, Naimah
building Tun Abdul Razak Library
collection Institutional Repository
content_provider Universiti Teknologi Mara
content_source UiTM Institutional Repository
continent Asia
country Malaysia
description Data depublication is method that help reduce the redundant data in storage capacity. With the rapid growth of digital data that generated in the digital world, the capacity of storage usage will increase rapidly. To achieve deduplication efficiency in system storage, the duplicate data need to be eliminated. To eliminated the duplicate data, the file unique value or hash value need to compare and the files that have the same hash value will be remove. This method basically will help to improve the storage capacity and efficiency. The hash value is generated by using hashing algorithm such as Message Digest 5 (MD5) and Secure Hashing Algorithm 1 (SHA-1). The hash functions should not create the same index value for the different data. If there is lack of analysis on the hashing algorithm, the deduplication technique cannot be improved for future research and the evolution of data deduplication can be slow because the performance metric for each hashing algorithm is not clear enough. The objective of this project is to compare MD5 & SHA-1 algorithm in data deduplication technique and to evaluate the MD5 & SHA-1 algorithm, length of message digest and speed using deduplication software. The simulation was conducted using File Alyzer, Clone Files Checker and AllDup software. The result of this simulation had been analysed based on three performance metrics which is efficiency, message digest length and the speed. There were two type of dataset which is video and document files with four different sizes. The time taken of the hashing algorithm generate the hash value were recorded. The findings in this project is the MD5 speed performance is better than SHA-1 hashing algorithm because it generates the hash value faster due to the length of message digest in MD5 is shorter than SHA-1. The recommendation for future work is to evaluate various type of data and different type of hashing algorithm.
format Thesis
id my.uitm.ir-26848
institution Universiti Teknologi Mara
language en
publishDate 2019
record_format eprints
spelling my.uitm.ir-268482019-12-10T07:18:52Z https://ir.uitm.edu.my/id/eprint/26848/ Data depublication using : Hashing algorithm / Naimah Nayan Nayan, Naimah Algorithms Data depublication is method that help reduce the redundant data in storage capacity. With the rapid growth of digital data that generated in the digital world, the capacity of storage usage will increase rapidly. To achieve deduplication efficiency in system storage, the duplicate data need to be eliminated. To eliminated the duplicate data, the file unique value or hash value need to compare and the files that have the same hash value will be remove. This method basically will help to improve the storage capacity and efficiency. The hash value is generated by using hashing algorithm such as Message Digest 5 (MD5) and Secure Hashing Algorithm 1 (SHA-1). The hash functions should not create the same index value for the different data. If there is lack of analysis on the hashing algorithm, the deduplication technique cannot be improved for future research and the evolution of data deduplication can be slow because the performance metric for each hashing algorithm is not clear enough. The objective of this project is to compare MD5 & SHA-1 algorithm in data deduplication technique and to evaluate the MD5 & SHA-1 algorithm, length of message digest and speed using deduplication software. The simulation was conducted using File Alyzer, Clone Files Checker and AllDup software. The result of this simulation had been analysed based on three performance metrics which is efficiency, message digest length and the speed. There were two type of dataset which is video and document files with four different sizes. The time taken of the hashing algorithm generate the hash value were recorded. The findings in this project is the MD5 speed performance is better than SHA-1 hashing algorithm because it generates the hash value faster due to the length of message digest in MD5 is shorter than SHA-1. The recommendation for future work is to evaluate various type of data and different type of hashing algorithm. 2019-12-12 Thesis NonPeerReviewed text en https://ir.uitm.edu.my/id/eprint/26848/1/TD_NAIMAH%20NAYAN%20CS%20R%2019_5.pdf Data depublication using : Hashing algorithm / Naimah Nayan. (2019) Degree thesis, thesis, Universiti Teknologi MARA, Perlis.
spellingShingle Algorithms
Nayan, Naimah
Data depublication using : Hashing algorithm / Naimah Nayan
title Data depublication using : Hashing algorithm / Naimah Nayan
title_full Data depublication using : Hashing algorithm / Naimah Nayan
title_fullStr Data depublication using : Hashing algorithm / Naimah Nayan
title_full_unstemmed Data depublication using : Hashing algorithm / Naimah Nayan
title_short Data depublication using : Hashing algorithm / Naimah Nayan
title_sort data depublication using : hashing algorithm / naimah nayan
topic Algorithms
url https://ir.uitm.edu.my/id/eprint/26848/1/TD_NAIMAH%20NAYAN%20CS%20R%2019_5.pdf
https://ir.uitm.edu.my/id/eprint/26848/
url_provider http://ir.uitm.edu.my/