Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 908 | 06-04-2025 11:36 PM | |
| 1509 | 03-23-2025 05:23 AM | |
| 744 | 03-17-2025 10:18 AM | |
| 2680 | 03-05-2025 01:34 PM | |
| 1786 | 03-03-2025 01:09 PM |
05-30-2018
03:56 PM
@Rodrigo Hjort Have you ever restarted your cluster after kerberization? The assumption here is your OS is RHEL/CentOS/Oracle Linux 6,if not please adapt the commands Can you share the contents of krb5.conf kdc.conf kadm5.acl Make sure the Kerberos daemons are up and running # kadmin status
# krb5kdc status If down # service kadmin start
# service krb5kdc start Make sure NTPD is running so the time in your cluster is in sync. To sum up, the ticket lifetime is the minimum of the following values: max_life in kdc.conf on the KDC servers. ticket_lifetime in krb5.conf on the client machine. maxlife for the user principal. Please revert
... View more
05-30-2018
02:34 PM
@Jyotsna Ganji Great it worked for you happy hadooping !!!!
... View more
05-30-2018
12:06 PM
@Jyotsna Ganji What OS version is that if RHEL/Centos7 can you try this # sudo mkdir -p /home/zookeeper Could you also check this # ll /home Please let me know
... View more
05-30-2018
11:42 AM
1 Kudo
@Jyotsna Ganji The error "No home directory for zookeeper" tells you all Can you do the following # mkdir -p /home/zookeeper Change ownership # chown zookeeper /home/zookeeper Then retry
... View more
05-29-2018
09:50 PM
@Samant Thakur The JMX response typically indicates 3 things why the datanode was not accessible. Network Issue DataNode is down Excessively long Garbage Collection This message comes from "/usr/lib/python2.6/site-packages/ambari_agent/alerts/metric_alert.py" script and following is the logic: if isinstance(self.metric_info, JmxMetric):
jmx_property_values, http_code = self._load_jmx(alert_uri.is_ssl_enabled, host, port, self.metric_info)
if not jmx_property_values and http_code in [200, 307]:
collect_result = self.RESULT_UNKNOWN
value_list.append('HTTP {0} response (metrics unavailable)'.format(str(http_code)))
elif not jmx_property_values and http_code not in [200, 307]:
raise Exception("[Alert][{0}] Unable to extract JSON from JMX response".format(self.get_name()))
else:
value_list.extend(jmx_property_values)
check_value = self.metric_info.calculate(value_list)
value_list.append(check_value) Network i MTU (Maximum Transmission Unit) is related to TCP/IP networking in Linux. It refers to the size (in bytes) of the largest datagram that a given layer of a communications protocol can pass at a time. It should be identical on all the nodes.MTU is set in /etc/sysconfig/network-scripts/ifcfg-ethx You can see current MTU setting with ifconfig command under Linux: $ netstat -i check the second row or $ ip link list - Check the host file on those failing nodes - Check if DNS server is having problems in name resolution. Run TestDFSIO performance tests yarn jar /usr/hdp/2.x.x.x.x/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-*tests.jar TestDFSIO -write -nrFiles 100 -fileSize 100 TestDFSIO Read Test hadoop jar Iperf Is a widely used tool for network performance measurement and tuning See Typical HDP Cluster Network Configuration Best Practices Datanode is down Restart the datanode using Ambari or manually Garbage Collection Running the GCViewer Enable GC logging for Datanode service
Open hadoop-env.sh and look for the following line export HADOOP_DATANODE_OPTS="-Dcom.sun.management.jmxremote -Xms2048m -Xmx2048m -Dhadoop.security.logger=ERROR,DRFAS $HADOOP_DATANODE_OPTS"
Insert the following into HADOOP_DATANODE_OPTS param -verbose:gc
-XX:+PrintGCDetails
-Xloggc:${HADOOP_LOG_DIR}/hadoop-hdfs-datanode-`date +'%Y%m%d%H%M'`.gclog
-XX:+UseGCLogFileRotation
-XX:NumberOfGCLogFiles=20
After adding the GC log pram the HADOOP_DATANODE_OPTS should look like this export HADOOP_DATANODE_OPTS="-Dcom.sun.management.jmxremote -Xms2048m -Xmx2048m -Dhadoop.security.logger=ERROR,DRFAS -verbose:gc -XX:+PrintGCDetails -Xloggc:${HADOOP_LOG_DIR}/hadoop-hdfs-datanode-`date +'%Y%m%d%H%M'`.gclog -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=20 $HADOOP_DATANODE_OPTS" The log should give you a detailed info. Hope that helps
... View more
05-29-2018
08:33 PM
@sudi ts To you have access to the GCP IAM console? When treating a service account as a resource, you can grant permission to a user to access that service account. You can grant the Owner, Editor, Viewer, or Service Account User role to a user to access the service account.
... View more
05-29-2018
06:56 PM
@Saurabh Srivastava Any updates?
... View more
05-29-2018
06:54 PM
@Kenneth K Any updates?
... View more
05-29-2018
06:43 PM
@sudi ts This is clearly a permission issue "Permission denied: user=hdpuser1, path="gs://bucket/":hive:hive:drwx------)" Have you tried using ACL's gsutil acl ch -u hdpuser1:WRITE gs://bucket/ And retry
... View more