Member since
08-08-2017
1652
Posts
30
Kudos Received
11
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2062 | 06-15-2020 05:23 AM | |
| 17119 | 01-30-2020 08:04 PM | |
| 2225 | 07-07-2019 09:06 PM | |
| 8583 | 01-27-2018 10:17 PM | |
| 4847 | 12-31-2017 10:12 PM |
06-16-2019
10:47 PM
second is it safe to delete the folder - hdfs dfs -rm -r /tmp/hive/hive
... View more
06-16-2019
10:46 PM
@dear jay - what is the meaning of hive.scratchdir.lock when is set to false?
... View more
06-16-2019
08:56 PM
hi all we have ambari cluster ( HDP version - 2.5.4 ) in the spark thrift log we can see the error about - /tmp/hive/hive is exceeded: limit=1048576 items=1048576 we try to delete the old files under /tmp/hive/hive , but there are a million of files and we cant delete them because hdfs dfs -ls /tmp/hive/hive isn't return any output any suggestion ? how to delete the old files in spite there are a million of files? or any other solution/? * for now spark thrift server isn't started successfully because this error , also hiveserver2 not started also Caused by: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.FSLimitException$MaxDirectoryItemsExceededException): The directory item limit of /tmp/hive/hive is exceeded: limit=1048576 items=1048576 at org.apache.hadoop.ipc.Server$Han
dler.run(Server.java:2347) second can we purge the files? by cron or other? hdfs dfs -ls /tmp/hive/hive
Found 4 items
drwx------ - hive hdfs 0 2019-06-16 21:58 /tmp/hive/hive/2f95f6a5-76ad-487e-968c-1873264a3a9c
drwx------ - hive hdfs 0 2019-06-16 21:45 /tmp/hive/hive/368d201c-cedf-48dc-bbad-f13d6aed7016
drwx------ - hive hdfs 0 2019-06-16 21:58 /tmp/hive/hive/717fb013-535b-4279-a12e-4fc4261c4d68
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
-
Apache Hive
06-16-2019
03:51 PM
from - https://stackoverflow.com/questions/44235019/delete-files-older-than-10days-on-hdfs
... View more
06-16-2019
03:49 PM
we try the following to remove files that older then 10 days , but because there are so many files then files not deleted at all hdfs dfs -ls /tmp/hive/hive | tr -s " " | cut -d' ' -f6-8 | grep "^[0-9]" | awk 'BEGIN{ MIN=14400; LAST=60*MIN; "date +%s" | getline NOW } { cmd="date -d'\''"$1" "$2"'\'' +%s"; cmd | getline WHEN; DIFF=NOW-WHEN; if(DIFF > LAST){ print "Deleting: "$3; system("hdfs dfs -rm -r "$3) }}'
... View more
06-16-2019
03:47 PM
is it safe to remove the /tmp/hive/hive folder? ( from hdfs ) as hdfs dfs -rm -r /tmp/hive/hive the reason for that because under /tmp/hive/hive we have thousand of files and we cant delete therm hdfs dfs -ls /tmp/hive/
Found 7 items
drwx------ - admin hdfs 0 2019-03-05 12:00 /tmp/hive/admin
drwx------ - drt hdfs 0 2019-06-16 14:02 /tmp/hive/drt
drwx------ - ambari-qa hdfs 0 2019-06-16 15:11 /tmp/hive/ambari-qa
drwx------ - anonymous hdfs 0 2019-06-16 08:57 /tmp/hive/anonymous
drwx------ - hdfs hdfs 0 2019-06-13 08:42 /tmp/hive/hdfs
drwx------ - hive hdfs 0 2019-06-13 10:58 /tmp/hive/hive
drwx------ - root hdfs 0 2018-07-17 23:37 /tmp/hive/root
You have mail in /var/spool/mail/root
... View more
Labels:
06-16-2019
05:32 AM
also I cant start the journal node ( on the bade namenode ) 2019-06-16 05:29:39,734 WARN namenode.FSImage (EditLogFileInputStream.java:scanEditLog(359)) - Caught exception after scanning through 0 ops from /hadoop/hdfs/journal/hdfsha/current/edits_inprogress_0000000000018783114 while determining its valid length. Position was 1032192
java.io.IOException: Can't scan a pre-transactional edit log.
at org.apache.hadoop.hdfs.server.namenode.FSEditLogOp$LegacyReader.scanOp(FSEditLogOp.java:4974)
at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.scanNextOp(EditLogFileInputStream.java:245)
at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.scanEditLog(EditLogFileInputStream.java:355)
at org.apache.hadoop.hdfs.server.namenode.FileJournalManager$EditLogFile.scanLog(FileJournalManager.java:551)
at org.apache.hadoop.hdfs.qjournal.server.Journal.scanStorageForLatestEdits(Journal.java:192)
at org.apache.hadoop.hdfs.qjournal.server.Journal.<init>(Journal.java:152)
at org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:90)
at org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:99)
at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.getJournalState(JournalNodeRpcServer.java:127)
at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.getJournalState(QJournalProtocolServerSideTranslatorPB.java:118)
at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25415)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2347)
2019-06-16 05:29:39,734 WARN namenode.FSImage (EditLogFileInputStream.java:scanEditLog(364)) - After resync, position is 1032192
... View more
06-16-2019
04:02 AM
@Geoffrey Shelton Okot no both namenode started as standby and then the namenode stooped
... View more
06-14-2019
06:44 AM
so according to this info do you recommended to set the value as is ( 31 ) or other ?
... View more
06-14-2019
06:43 AM
on the good name node we have the number 31 on the bad name node we have also the number 31 on the other journal node we have the number 28
... View more