Member since
06-29-2015
47
Posts
8
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1175 | 02-08-2021 08:52 AM | |
1312 | 03-16-2017 04:52 PM |
05-18-2021
08:00 AM
I was able to find out number of mappers used by distcp using below command: MAPPERS=`yarn container -list $app | grep 'Total number of containers' | awk -F: '{print $2}'` Next step is to only look for distcp job which is doing a copy from/to hdfs (and not to S3). What's the best way to get around it?
... View more
05-06-2021
02:08 PM
distcp - How to determine number of mappers used by distcp job(s) at cluster level? Sometime we run into network bandwidth issue caused by distcp job(s) running too many mappers or too many distcp jobs. Our plan is to trigger DataDog alert when the total number of mappers used by distcp jobs (at cluster level) reach at defined number (ex: 100). We are open to explore the "-bandwidth" option. We have many users who will be submitting a job from diff edge nodes. so, we don't want to use the "ps" command at server level. Please help us rectify the issue. Thanks in advance.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
HDFS
02-08-2021
08:52 AM
Sorry, I forgot to post earlier. I was able to fix my own issue. I had to restart Ranger and Solr. One of the Solr instance failed on restart. But I was able to see Audit tab and other setting on Ranger. Thanks for looking into it.
... View more
02-05-2021
12:49 PM
The "Audit" tab is missing from Ranger UI. It was working fine before. We have not made any changes to the cluster. I'm unable to find any article regarding this issue. I'm hoping go them some help from the community. Thanks. The audit logs are working coming to HDFS just fine. [hdfs@mynode~]$ hdfs dfs -ls /ranger/audit/
Found 7 items
drwx------ - hdfs hdfs 0 2021-02-05 00:00 /ranger/audit/hdfs
drwx------ - hive hive 0 2019-05-22 15:18 /ranger/audit/hive2
drwx------ - hive hive 0 2021-02-05 00:00 /ranger/audit/hiveServer2
drwxr-x--- - kms kms 0 2021-02-05 00:00 /ranger/audit/kms
drwx------ - knox knox 0 2017-01-19 15:48 /ranger/audit/knox
drwxr-x--- - nifi nifi 0 2021-02-04 18:00 /ranger/audit/nifi
drwx------ - yarn yarn 0 2021-02-05 00:00 /ranger/audit/yarn Components version: HDP: 2.6.5.0
Ranger: 0.7.0
Ambari Infra: 0.1.0 Ranger Audit Config:
... View more
Labels:
- Labels:
-
Apache Ranger
10-28-2019
12:19 PM
I'm getting the same error. This is the response I received form Cloudera support "Only the dfs commands such as ls/put/mv etc works on wasb using the wasb connector. Admin commands such as dfsadmin as well fsck works only with native hadoop/hdfs implementation"
... View more
04-26-2019
02:51 PM
Do you have latest recommendations? Most of our hadoop processing is on Hive/Tez and Spark.
... View more
10-12-2017
07:02 PM
@Sandeep More I logged in and ran kinit. So have valid ticket and able to run other hdfs commands.
... View more
10-10-2017
08:55 PM
@Sandeep More I'm getting below error on HDP 2.6.1, please see if you can help me. I'm able to browse "webhdfs" without knox. curl -k -i --negotiate -u : https://myhost.mydomain:8443/gateway/default/webhdfs/v1/tmp?op=LISTSTATUS
HTTP/1.1 401 Authentication required
Date: Tue, 10 Oct 2017 20:41:13 GMT
WWW-Authenticate: Negotiate
Set-Cookie: hadoop.auth=; Path=gateway/default; Domain=mydomain.com; Secure; HttpOnly
Content-Type: text/html; charset=ISO-8859-1
Cache-Control: must-revalidate,no-cache,no-store
Content-Length: 320
Server: Jetty(9.2.15.v20160210)
HTTP/1.1 403 org.apache.hadoop.security.authentication.client.AuthenticationException
Date: Tue, 10 Oct 2017 20:41:14 GMT
Set-Cookie: hadoop.auth=; Path=gateway/default; Domain=mydomain.com; Secure; HttpOnly
Content-Type: text/html; charset=ISO-8859-1
Cache-Control: must-revalidate,no-cache,no-store
Content-Length: 314
Server: Jetty(9.2.15.v20160210)
<html>
<head>
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8"/>
<title>Error 403 Forbidden</title>
</head>
<body><h2>HTTP ERROR 403</h2>
<p>Problem accessing /gateway/default/webhdfs/v1/tmp. Reason:
<pre> Forbidden</pre></p><hr><i><small>Powered by Jetty://</small></i><hr/>
</body>
</html><br> <provider>
<role>authentication</role>
<name>HadoopAuth</name>
<enabled>true</enabled>
<param>
<name>config.prefix</name>
<value>hadoop.auth.config</value>
</param>
<param>
<name>hadoop.auth.config.signature.secret</name>
<value>/etc/security/http_secret</value>
</param>
<param>
<name>hadoop.auth.config.type</name>
<value>kerberos</value>
</param>
<param>
<name>hadoop.auth.config.simple.anonymous.allowed</name>
<value>false</value>
</param>
<param>
<name>hadoop.auth.config.token.validity</name>
<value>1800</value>
</param>
<param>
<name>hadoop.auth.config.cookie.domain</name>
<value>mydomain.com</value>
</param>
<param>
<name>hadoop.auth.config.cookie.path</name>
<value>gateway/default</value>
</param>
<param>
<name>hadoop.auth.config.kerberos.principal</name>
<value>HTTP/_HOST@TEST.COM</value>
</param>
<param>
<name>hadoop.auth.config.kerberos.keytab</name>
<value>/etc/security/keytabs/spnego.service.keytab</value>
</param>
<param>
<name>hadoop.auth.config.kerberos.name.rules</name>
<value>DEFAULT</value>
</param>
</provider><br>
... View more
04-13-2017
06:01 PM
I'm facing the same issue: YARN:
Memory allocated for all YARN containers on a node = 16G
Minimum Container Size (Memory) = 2G
Maximum Container Size (VCores) = 3
Hive:
% of cluster capacity = 40%
Memory per daemon = 8192
Number of LLAP Daemons = 1
Memory per daemon = 8192
In-Memory Cache per Daemon = 2048
Maximum CPUs per Daemon = 3
I do see this error messages on the RM UI: Diagnostics: Unstable Application Instance : - failed with
component LLAP failed 'recently' 6 times (4 in startup); threshold is 5 - last failure: Failure container_e29_1492031103210_0001_01_000007 on host host1.fqdn (0): http://host1.fqdn:19888/jobhistory/logs/host1.fqdn:45454/container_e29_1492031103210_0001_01_000007/ctx/hive
... View more
04-10-2017
08:51 PM
@Neeraj Sabharwal Please see if you can help me.
... View more