How would codecs useful to Hadoop?

Viewing 1 reply thread
  • Author
    Posts
    • #6258
      DataFlair TeamDataFlair Team
      Spectator

      Explain how codecs will be useful to Apache Hadoop?

    • #6259
      DataFlair TeamDataFlair Team
      Spectator

      By default, the security of Hadoop deals with large volumes of data. Surely, compression decompression(codecs) saves great deal of storage space as well as speed up the movement of data through cluster.
      It can also be useful to compress the intermediate output of the map phase. Because MapReduce output is written to disk and then shipped across the reduce tasks compressing the size of this data can lead to great level of performance improvements.

      A codec is the implementation of a compression-decompression algorithm. In Hadoop, a codec is represented by an implementation of the CompressionCodec interface. So, for example, GzipCodec encapsulates the compression and decompression algorithm for gzip.

Viewing 1 reply thread
  • You must be logged in to reply to this topic.