linux - Hadoop No space left on device erro when there is space available -
i have 5 linux machines cluster. there 3 data nodes , 1 master. @ 50% hdfs storage available on each data nodes. run mapreduce job, failed following error
2017-08-21 17:58:47,627 warn org.apache.hadoop.hdfs.dfsclient: error recovery blk_6835454799524976171_3615612 bad datanode[0] 10.11.1.42:50010 2017-08-21 17:58:47,628 warn org.apache.hadoop.hdfs.dfsclient: error recovery block blk_6835454799524976171_3615612 in pipeline 10.11.1.42:50010, 10.11.1.43:50010: bad datanode 10.11.1.42:50010 2017-08-21 17:58:51,785 error org.apache.hadoop.mapred.child: error in synclogs: java.io.ioexception: no space left on device
while on each system df -h
gives following information
filesystem size used avail use% mounted on devtmpfs 5.9g 0 5.9g 0% /dev tmpfs 5.9g 84k 5.9g 1% /dev/shm tmpfs 5.9g 9.1m 5.9g 1% /run tmpfs 5.9g 0 5.9g 0% /sys/fs/cgroup /dev/mapper/centos-root 50g 6.8g 44g 14% / /dev/sdb 1.8t 535g 1.2t 31% /mnt/11fd6fcc-1f87-4f1e-a53c-54cc7117759c /dev/mapper/centos-home 412g 155g 59m 100% /home /dev/sda1 494m 348m 147m 71% /boot tmpfs 1.2g 16k 1.2g 1% /run/user/42 tmpfs 1.2g 0 1.2g 0% /run/user/1000
as clear above sdb dicsk (sdd) 31% used centos-home 100%. while hadoop using local file system in mapreduce job when there enough hdfs available? problem? have search @ google , found many such problem no 1 covers situation.
synclogs
not use hdfs, writes hadoop.log.dir
if you're using mapreduce, value of hadoop.log.dir
in /etc/hadoop/conf/taskcontroller.cfg.
if you're using yarn, value of yarn.nodemanager.log-dirs
in yarn-site.xml.
one of these should point you're writing logs. once figure out filesystem has problem, can free space there.
another thing remember "no space left on device" if you've exhausted inodes on disk. df -i
show this.
wiki
Comments
Post a Comment