DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION
This paper explains Data Ingestion which is the process of collecting data. Data ingestion usually occurs in the internal organization so that the organization can analyze the data further. A famous file storage for big data analysis is Hadoop Distributed File System (HDFS). There are two tools r...
Saved in:
Main Author: | |
---|---|
Format: | Final Year Project |
Language: | English |
Published: |
IRC
2019
|
Subjects: | |
Online Access: | http://utpedia.utp.edu.my/20909/1/LAI%20ZHEN%20YEAN_22888.pdf http://utpedia.utp.edu.my/20909/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my-utp-utpedia.20909 |
---|---|
record_format |
eprints |
spelling |
my-utp-utpedia.209092021-09-09T20:08:43Z http://utpedia.utp.edu.my/20909/ DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION Lai, Zhen Yean Q Science (General) This paper explains Data Ingestion which is the process of collecting data. Data ingestion usually occurs in the internal organization so that the organization can analyze the data further. A famous file storage for big data analysis is Hadoop Distributed File System (HDFS). There are two tools related to data ingestion in Hadoop, which are Apache Sqoop and Apache Flume. Apache Sqoop is a tool to transfer data between Hadoop and Relational Database Management System (RDBMS) . Apache Flume is a distributed service to collect data from multiple variety of sources and forward to Hadoop Storage. The concerns of these tools are they do not have built-in data compression and data hiding feature during the data transmission. The proposed solution to this concern is applying the Fixed Length Coding (FLC) compression with Audio Steganography technique by using a new data ingestion method to achieve data compression and data hiding. The proposed solution methodology is implementing the data compression and audio steganography during the transmission of the data from RDBMS to Hadoop Distributed File System (HDFS) Storage. However, there is an inefficient aspect which is the capability of overcome data loss during audio steganography. Further performance evaluation is performed to valid the data transmission, the evaluation parameters including compression ratio, signal to noise ratio and information loss. IRC 2019-09 Final Year Project NonPeerReviewed application/pdf en http://utpedia.utp.edu.my/20909/1/LAI%20ZHEN%20YEAN_22888.pdf Lai, Zhen Yean (2019) DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION. IRC, Universiti Teknologi PETRONAS. (Submitted) |
institution |
Universiti Teknologi Petronas |
building |
UTP Resource Centre |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Teknologi Petronas |
content_source |
UTP Electronic and Digitized Intellectual Asset |
url_provider |
http://utpedia.utp.edu.my/ |
language |
English |
topic |
Q Science (General) |
spellingShingle |
Q Science (General) Lai, Zhen Yean DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
description |
This paper explains Data Ingestion which is the process of collecting data. Data ingestion usually
occurs in the internal organization so that the organization can analyze the data further. A
famous file storage for big data analysis is Hadoop Distributed File System (HDFS). There are
two tools related to data ingestion in Hadoop, which are Apache Sqoop and Apache Flume.
Apache Sqoop is a tool to transfer data between Hadoop and Relational Database Management
System (RDBMS) . Apache Flume is a distributed service to collect data from multiple variety of
sources and forward to Hadoop Storage. The concerns of these tools are they do not have built-in
data compression and data hiding feature during the data transmission. The proposed solution to
this concern is applying the Fixed Length Coding (FLC) compression with Audio
Steganography technique by using a new data ingestion method to achieve data compression
and data hiding. The proposed solution methodology is implementing the data compression and
audio steganography during the transmission of the data from RDBMS to Hadoop Distributed
File System (HDFS) Storage. However, there is an inefficient aspect which is the capability of
overcome data loss during audio steganography. Further performance evaluation is performed to
valid the data transmission, the evaluation parameters including compression ratio, signal to
noise ratio and information loss. |
format |
Final Year Project |
author |
Lai, Zhen Yean |
author_facet |
Lai, Zhen Yean |
author_sort |
Lai, Zhen Yean |
title |
DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
title_short |
DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
title_full |
DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
title_fullStr |
DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
title_full_unstemmed |
DATA COMPRESSION AND DATA HIDING DURING LARGE DATA INGESTION |
title_sort |
data compression and data hiding during large data ingestion |
publisher |
IRC |
publishDate |
2019 |
url |
http://utpedia.utp.edu.my/20909/1/LAI%20ZHEN%20YEAN_22888.pdf http://utpedia.utp.edu.my/20909/ |
_version_ |
1739832810976313344 |
score |
13.160551 |