A Data Compression Approach for Big Data Classification
Eskadmas Ayenew Tefera
A Data Compression Approach for Big Data Classification.
Rel. Paolo Garza. Politecnico di Torino, Master of science program in Computer Engineering, 2020
|
Preview |
PDF (Tesi_di_laurea)
- Thesis
Licence: Creative Commons Attribution Non-commercial No Derivatives. Download (1MB) | Preview |
Abstract
Abstract - A data compression approach for big data classification is used, first, to compress large size dataset into small size dataset, then to created and build the classification model for the compressed dataset and evaluate the model’s accuracy. Data compression algorithms are used to compress and reduce the size of dataset. The compression mechanisms help to increase and optimize operational efficiencies, enable cost reductions, and reduce risks for the business operations. It is becoming costly to process large size datasets, which need to be reduced by using compression techniques. The compression algorithm controls each bit of a dataset and optimizes the size without losing any data subsequently by using a lossless data compression approach.
In a lossless data compression technique, data can be compressed without loss
Relators
Publication type
URI
![]() |
Modify record (reserved for operators) |
