Member since
06-18-2020
3
Posts
0
Kudos Received
0
Solutions
06-19-2020
07:48 AM
Hi. Tis hadoop cluster hacer 1.4PB size, so for this node we have this situation size on the Mount points: [root@ithbda108 ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/md2 459G 50G 385G 12% / tmpfs 126G 36K 126G 1% /dev/shm /dev/md0 453M 77M 349M 19% /boot /dev/sda4 6.6T 6.3T 323G 96% /u01 /dev/sdb4 6.6T 6.3T 321G 96% /u02 /dev/sdc1 7.1T 6.8T 314G 96% /u03 /dev/sdd1 7.1T 6.8T 314G 96% /u04 /dev/sde1 7.1T 6.8T 318G 96% /u05 /dev/sdf1 7.1T 6.8T 323G 96% /u06 /dev/sdg1 7.1T 6.8T 325G 96% /u07 /dev/sdh1 7.1T 6.8T 323G 96% /u08 /dev/sdi1 7.1T 6.8T 324G 96% /u09 /dev/sdj1 7.1T 6.8T 324G 96% /u10 /dev/sdk1 7.1T 6.8T 324G 96% /u11 /dev/sdl1 7.1T 6.8T 322G 96% /u12 cm_processes 126G 200M 126G 1% /var/run/cloudera-scm-agent/process ithbda103.sopbda.telcel.com:/opt/exportdir 459G 338G 98G 78% /opt/shareddir I suppose that it can be an issue about space disk and there is no space left on the device at the time of writing into log4j. Any idea in what action can we do to solve the space left on the mount points? Some cloudera procedure to optimize that the process can be up?
... View more
06-18-2020
07:59 PM
I checked all the file system with the command "du -sk" and the response was ok. I tried to restart again the Datanode and now the error is another: Successfully obtained privileged resources (streaming port = ServerSocket[addr=/0.0.0.0,localport=1004] ) (http listener port = 1006) Opened info server at /0.0.0.0:1006 Starting regular datanode initialization log4j:ERROR Failed to flush writer, java.io.IOException: No space left on device at java.io.FileOutputStream.writeBytes(Native Method) at java.io.FileOutputStream.write(FileOutputStream.java:326) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.RollingFileAppender.subAppend(RollingFileAppender.java:276) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.apache.commons.logging.impl.Log4JLogger.info(Log4JLogger.java:176) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.releaseShortCircuitFds(DataXceiver.java:408) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReleaseShortCircuitFds(Receiver.java:236) at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:124) at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:246) at java.lang.Thread.run(Thread.java:748) Service exit with a return value of 143 Initializing secure datanode resources Opened streaming server at /0.0.0.0:1004 Successfully obtained privileged resources (streaming port = ServerSocket[addr=/0.0.0.0,localport=1004] ) (http listener port = 1006) Opened info server at /0.0.0.0:1006 Starting regular datanode initialization Service exit with a return value of 143 Initializing secure datanode resources I checked the ports LISTEN on the node to check the socket 1006 and there is no process running: [root@ithbda106 ~]# netstat -netupa | egrep "1004|1006" | grep LIST <<< No return >> How can I get the solution for this?
... View more
06-18-2020
05:59 PM
Hi, After a reboot, I have got this error to try to restart a datanode :"Could not get disk usage information". Any idea in how to solve this problem?
... View more
Labels:
- Labels:
-
HDFS