How would codecs useful to Hadoop?
-
-
Explain how codecs will be useful to Apache Hadoop?
-
By default, the security of Hadoop deals with large volumes of data. Surely, compression decompression(codecs) saves great deal of storage space as well as speed up the movement of data through cluster.
It can also be useful to compress the intermediate output of the map phase. Because MapReduce output is written to disk and then shipped across the reduce tasks compressing the size of this data can lead to great level of performance improvements.
A codec is the implementation of a compression-decompression algorithm. In Hadoop, a codec is represented by an implementation of the CompressionCodec interface. So, for example, GzipCodec encapsulates the compression and decompression algorithm for gzip.
- You must be logged in to reply to this topic.