Member since
03-14-2016
4721
Posts
1111
Kudos Received
874
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2445 | 04-27-2020 03:48 AM | |
4882 | 04-26-2020 06:18 PM | |
3976 | 04-26-2020 06:05 PM | |
3219 | 04-13-2020 08:53 PM | |
4925 | 03-31-2020 02:10 AM |
06-16-2019
11:07 PM
@Michael Bronson For now we can enable "hive.server2.clear.dangling.scratchdir=true" for HiveServer2 via Custom hive-site.
... View more
06-16-2019
11:02 PM
1 Kudo
@Michael Bronson I do not remember/think of any specific idem to check at this point, But as long as you are able to run your Hive Queries without any issue and HiveService checks are also running fine. I think we should be good.
... View more
06-16-2019
10:59 PM
@Michael Bronson You can add those scratchdir parameters inside the in custom hive-site.xml via Ambari UI by clicking the Add Property option.
... View more
06-16-2019
10:54 PM
1 Kudo
@Michael Bronson To clean up the Hive scratch directory manually may not be a safe option for a multi-user environment (where multiple users might be executing the hive queries concurrently) since it will accidentally remove the scratch directory in use.
... View more
06-16-2019
10:51 PM
1 Kudo
@Michael Bronson "hive.scratchdir.lock" : When true, holds a lock file in the scratch directory. If a Hive process dies and accidentally leaves a dangling scratchdir behind, the cleardanglingscratchdir tool will remove it. When false, does not create a lock file and therefore the cleardanglingscratchdir tool cannot remove any dangling scratch directories. Regarding your query "second is it safe to delete the folder - /tmp/hive/hive" >>> I do not think that we should do it on our own. As the whole purpose of the following JIRA was to introduce some tool like "cleardanglingscratchdir" to safely remove the scratch contents. https://issues.apache.org/jira/browse/HIVE-13429 .
... View more
06-16-2019
10:32 PM
1 Kudo
@Michael Bronson As an immediate turnaround may be you would like to double / increase the value set for the following HDFS parameter "dfs.namenode.fs-limits.max-directory-items". It Defines the maximum number of items that a directory may contain. Cannot set the property to a value less than 1 or more than 6400000. (default value : 1048576) Go to Ambari -> HDFS -> Configs -> Advanced -> Custom hdfs-site and add the key (dfs.namenode.fs-limits.max-directory-item As per the apache hive docs there seems to be some parameters and tools available to deal with such issue. Although i have not personally tested those tools. But looks like they were introduced to deal with similar issue long back as part of https://issues.apache.org/jira/browse/HIVE-13429 For example i see that the Hive Config "hive.exec.scratchdir" points to the "/tmp/hive" dir. Can you please check and let us know what is the value set for the following parameter "hive.scratchdir.lock". (if not set then default value will be "false"? Additionally you might want to refer about "hive.server2.clear.dangling.scratchdir" and "hive.start.cleanup.scratchdir" parameters of Hive Server config. Please refer to [1] the following link to know more about those parameters. There is a tool "cleardanglingscratchdir" mentioned as part of the link [2] may be you would like to read more about it. # hive --service cleardanglingscratchdir [-r] [-v] [-s scratchdir]
-r dry-run mode, which produces a list on console
-v verbose mode, which prints extra debugging information
-s if you are using non-standard scratch directory . [1] https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.scratchdir.lock. [2] https://cwiki.apache.org/confluence/display/Hive/Setting+Up+HiveServer2#SettingUpHiveServer2-ClearDanglingScratchDirTool
... View more
06-16-2019
10:31 PM
1 Kudo
@Michael Bronson As per the apache hive docs there seems to be some parameters and tools available to deal with such issue. Although i have not personally tested those tools. But looks like they were introduced to deal with similar issue long back as part of https://issues.apache.org/jira/browse/HIVE-13429 For example i see that the Hive Config "hive.exec.scratchdir" points to the "/tmp/hive" dir. Can you please check and let us know what is the value set for the following parameter "hive.scratchdir.lock". (if not set then default value will be "false"? Additionally you might want to refer about "hive.server2.clear.dangling.scratchdir" and "hive.start.cleanup.scratchdir" parameters of Hive Server config. Please refer to [1] the following link to know more about those parameters. There is a tool "cleardanglingscratchdir" mentioned as part of the link [2] may be you would like to read more about it. # hive --service cleardanglingscratchdir [-r] [-v] [-s scratchdir]
-r dry-run mode, which produces a list on console
-v verbose mode, which prints extra debugging information
-s if you are using non-standard scratch directory . [1] https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.scratchdir.lock. [2] https://cwiki.apache.org/confluence/display/Hive/Setting+Up+HiveServer2#SettingUpHiveServer2-ClearDanglingScratchDirTool
... View more
06-14-2019
08:24 AM
@Adil BAKKOURI Your Zookeeper process command output shows the log name something like following: root@RHBigData1:~# ps -ef | grep -i zookeeper
.
.
-Dzookeeper.log.file=zookeeper-zookeeper-server-RHBigData1.log . I doubt that your Zookeeper Hostname might not be correct. Usually the zookeeper log file name is generated based on the FQDN of the node. So if your Zookeeper FQDN is correctly set then it should be showing something like following: # hostname -f
master.rh.bigdata.cluster . So please verify of your Zookeeper host has the FQDN setup correctly and then restart the Zookeeper after fixing the FQDN. Later you should see the zookeeper log file name something like "zookeeper-zookeeper-server-master.rh.bigdata.cluster.log"
... View more
06-14-2019
02:36 AM
1 Kudo
@Gulshan Agivetova May be you can use the "InvokeHTTP" --> "ReplaceText" --> "PutFile" processors like following: Here in "ReplaceText" processor can be used something like following to replace Search Value = TAB with COMMA ReplaceText Properties: Search Value =
Replacement Value = ,
Character Set = UTF-8
Maximum Buffer Size = 1 MB
Replacement Strategy = Regex Replace
Evaluation Mode = Line-by-Line And then it should be able to replace the TABs with Comma. The output of the given inout should be something like: RELM,Satieva_microloans,6,3369,2019-06-12,11:42:00,772103971
RELM,Satieva_microloans,6,3369,2019-06-12,11:42:00,772104117
RELM,Satieva_microloans,6,3369,2019-06-12,11:42:00,772101000 .
... View more
06-13-2019
09:36 PM
@Adil BAKKOURI Ambari Simply makes use of the Zookeeper Smoke Test "zkSmoke.sh" script to verify the Zookeeper Connection and if it can create & delete the dummy ZNode "/zk_smoketest" inside the zookeeper host or not? As you see the following message: ExecutionFailed: Execution of '/var/lib/ambari-agent/tmp/zkSmoke.sh /usr/hdp/current/zookeeper-client/bin/zkCli.sh ambari-qa /usr/hdp/current/zookeeper-client/conf 2181 False kinit no_keytab no_principal /var/lib/ambari-agent/tmp/zkSmoke.out' returned 4. zk_node1=master.rh.bigdata.cluster
.
.
Exception in thread "main" org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /zk_smoketest Which means your Zookeeper process might not be running file. Most probably the Zookeeper port might not be accessible on host "master.rh.bigdata.cluster" So you will need to first check if your Zookeeper is running fine or not on host "master.rh.bigdata.cluster" and if the Zookeeper port 2181 is listening fine or not? # ps -ef | grep -i zookeeper
# netstat -tnlpa | grep 2181
# service iptables stop Most Possible Cause: If you see that Zookeeper process is not running then please try to restart it and check the Zookeper Logs to find any issue/ startup failure and please share the log. Most probably once your Zookeeper is running fine and the port 2181 is accessible then your Zookeeper Check Should also run fine. Firewall or Port Blocking issue: So please go to the host where the SmokeTest is failing and then try to telnet to Zookeeper Host & Port to see if that is accessible? # telnet master.rh.bigdata.cluster 2181
(OR)
# nc -v master.rh.bigdata.cluster 2181 Manual Test: You your self can try running the same command to verify this Zookeeper Zmoke Test run Try running this command on different Zookeeper nodes as it is. # /var/lib/ambari-agent/tmp/zkSmoke.sh /usr/hdp/current/zookeeper-client/bin/zkCli.sh ambari-qa /usr/hdp/current/zookeeper-client/conf 2181 False kinit no_keytab no_principal /var/lib/ambari-agent/tmp/zkSmoke.out . .
... View more