Member since
08-01-2013
187
Posts
10
Kudos Received
8
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2318 | 09-01-2016 09:26 PM | |
2005 | 10-14-2015 07:31 AM | |
2148 | 06-21-2015 06:02 PM | |
3858 | 02-26-2015 04:36 PM | |
4364 | 02-18-2015 12:18 AM |
04-27-2022
06:13 AM
Hi @zero , Did you try out commenting the renew_lifetime parameter in /etc/krb5.conf. I think I too had the same issue and this resolved my error. Mine was a CDP 7.1.4 cluster, but have faced similar issues in HDP also. https://community.cloudera.com/t5/Support-Questions/HDF-Zookeeper-server-not-starting-after-MIT-Kerberos-is/m-p/291367 Thanks, Vivek
... View more
07-31-2019
02:07 PM
@sparkd, While we can't be sure, it is likely that some permissions were changed on the /tmp directory so that the Service Monitor (that executes the HDFS canary health check) could not access the directory. Service Monitor utilizes the "hue" user and principal to access other resources so it is reasonable to assume that /tmp in HDFS did not allow the hue user or group to write to /tmp. Are you having similar trouble? If so, check your service monitor log file for stack traces and errors related to the hdfs canary.
... View more
06-01-2018
01:31 PM
@vvinaga It looks like it cannot talk to the HDFS NameNode from the logs. Could you check if HDFS is configured correctly to use Kerberos?
... View more
09-05-2017
07:04 AM
2 Kudos
Symptoms A Spark job fails with INTERNAL_FAILURE. In the WA (Workload Analytics) page of the job that failed, the following message is reported: org.apache.spark.SparkException: Application application_1503474791091_0002 finished with failed status
Diagnosis As the Telemetry Publisher didn't retrieve the application log due to a known bug, we have to diagnose the application logs (application_1503474791091_0002) directly, which are stored in the user's S3 bucket. If the following exception is found, it indicates that the application failed to resolve a dependency in the Hadoop class path: 17/08/24 13:13:33 INFO ApplicationMaster: Preparing Local resources Exception in thread "main" java.lang.NoSuchMethodError: org.apache.hadoop.tracing.TraceUtils.wrapHadoopConf(Ljava/lang/String;Lorg/apache/hadoop/conf/Configuration;)Lorg/apache/htrace/core/HTraceConfiguration; at org.apache.hadoop.fs.FsTracer.get(FsTracer.java:42) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:687) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:671) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:155) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2653) This most likely occurred because the jar may have been built using the another Hadoop distribution's repository, for example EMR (Amazon Elastic MapReduce)
Solution To resolve the issue, rebuild the application using the CDH repository, https://repository.cloudera.com/artifactory/cloudera-repos/, using Maven or sbt. The example of using Maven is as follows. https://www.cloudera.com/documentation/enterprise/release-notes/topics/cdh_vd_cdh5_maven_repo.html
... View more
06-21-2017
02:11 PM
3 Kudos
Question Where does Workload Analytics ingest user workloads into and analyze it?
Answer Workload Analytics runs as part of Cloudera Altus functionality on an environment operated by Cloudera. Telemetry Publisher, part of Cloudera Manager installation, sends user's workload to the environment as soon as a job ends and analyze it. Thereon the result shows up in the Cloudera Altus UI. https://www.cloudera.com/documentation/altus/topics/wa_overview.html
... View more
Labels:
06-21-2017
12:57 PM
1 Kudo
Question Is it possible to tune the thresholds for Health Check which Workload Analytics does based on user's requirements?
Answer No, it's not possible to tune the thresholds. Health Check uses predefined thresholds which are described in the document: https://www.cloudera.com/documentation/altus/topics/wa_analyze_jobs.html
... View more
02-06-2017
01:58 AM
hi, can u please tell me what changes did u make in th sudoers file? thnks
... View more
09-09-2016
09:34 AM
Hi, I did install from scratch a new cluster using m4 instance type and I could not reproduce the error. Thanks.
... View more
09-01-2016
09:26 PM
1 Kudo
Such custom jar file that HBase uses cannot be distributed across the hosts by CM automatically. You have to locate by yourself. HTH. -- Sent from my mobile
... View more