Update in HDFS

Discussion in 'Big Data and Analytics' started by _42734, Oct 27, 2018.

  1. _42734

    _42734 New Member

    Joined:
    Oct 4, 2018
    Messages:
    1
    Likes Received:
    0
    Can anybody explain how a update in an existing file is managed in a HDFS. I know update will be considered a new write request and new blocks will be written possibly in a new location... then what will happen with old blocks, will they be tagged as available in datanode block report to NN.
     
    #1
  2. Neha_Pandey

    Neha_Pandey Well-Known Member
    Simplilearn Support Alumni

    Joined:
    Jun 7, 2018
    Messages:
    65
    Likes Received:
    0
    Hi Learner,
    HDFS is designed to reliably store very large files across machines in a large cluster. It stores each file as a sequence of blocks; all blocks in a file except the last blockare the same size. The blocks of a file are replicated for fault tolerance. The blocksize and replication factor are configurable per file.
    You can delete the old files manually.

    Regards,
    Neha Pandey
     
    #2

Share This Page