Member since
04-20-2016
61
Posts
17
Kudos Received
13
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
928 | 06-30-2017 01:16 PM | |
934 | 06-30-2017 01:03 PM | |
1207 | 06-30-2017 12:50 PM | |
1178 | 06-30-2017 12:40 PM | |
13298 | 06-30-2017 12:36 PM |
01-05-2017
03:20 PM
Hi @yong yang. The SPARK_MAJOR_VERSION environment variable can be set by any user who logs on to a client machine to run Spark. The scope of the environment variable is local to the user session. Maybe in later version we may get option to to do this from ambari. Please do accept the answer, so it may useful to others too. Thx
... View more
01-05-2017
02:26 PM
3 Kudos
Hi @yong yang
By default, if more than one version of Spark is installed on a node, your job runs with the default version for your HDP package. The default version for HDP 2.5.0 is Spark 1.6.2. If more than one version of Spark is installed on a node, you can select which version of Spark runs your job. To do this, set the SPARK_MAJOR_VERSION environment variable to the desired version before you launch the job. Here is an example for a user who submits jobs using spark-submit under /usr/bin :
Navigate to a host where Spark 2.0 is installed. Change to the Spark2 client directory: cd /usr/hdp/current/spark2-client/ Set the SPARK_MAJOR_VERSION environment variable to 2: export SPARK_MAJOR_VERSION=2 Run the Spark Pi example: ./bin/spark-submit --class org.apache.spark.examples.SparkPi --master yarn-cluster --num-executors 1 --driver-memory 512m --executor-memory 512m --executor-cores 1 examples/jars/spark-examples*.jar 10
... View more
01-04-2017
01:07 PM
2 Kudos
Was an interesting issue faced last week. Putting here for bigger audience, might be helpful to others too. PROBLEM On one of the node, datanode and nodemanager were not coming up. Below is the error after starting from ambari. resource_management.core.exceptions.Fail: Execution of 'ambari-sudo.sh su hdfs -l -s /bin/bash -c 'ulimit -c unlimited ; /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config /usr/hdp/current/hadoop-client/conf start datanode'' returned 1. starting datanode, logging to /var/log/hadoop/hdfs/hadoop-hdfs-datanode-ny-node3.hwxblr.com.out
Error: Could not find or load main class org.apache.hadoop.hdfs.server.datanode.DataNode As datanode process itself wasn't loaded, so nothing was printed in datanode logs. Only thing we see in .out file is Error: Could not find or load main class org.apache.hadoop.hdfs.server.datanode.DataNode Verified in the jar, DataNode class was present. /usr/jdk64/jdk1.8.0_77/bin/jar -tvf /usr/hdp/2.5.0.0-1245/hadoop-hdfs/hadoop-hdfs-2.7.3.2.5.0.0-1245.jar | grep DataNode.class
org/apache/hadoop/hdfs/server/datanode/DataNode.class ROOT CAUSE @nvadivelu came to rescue. We used below utility to figure out which class was missing. public class Sample {
public static void main(String[] args) {
try {
org.apache.hadoop.hdfs.server.datanode.DataNode.main(args);
} catch (Throwable ex) {
ex.printStackTrace();
}
}
} We ran the above code, and it printed the exact class which wasn't able to load. /usr/jdk64/jdk1.8.0_77/bin/javac -cp `hadoop classpath` Sample.java
Sample.java:5: error: cannot access TraceAdminProtocol
org.apache.hadoop.hdfs.server.datanode.DataNode.main(args);
^
class file for org.apache.hadoop.tracing.TraceAdminProtocol not found
1 error TraceAdminProtocol clas is present hadoop-common jar. We grep this class in the hadoop-common jar, we didn't find. But on other host, where datanode was running fine, we got below results. grep "TraceAdminProtocol" /usr/hdp/2.5.0.0-1245/hadoop/hadoop-common-2.7.3.2.5.0.0-1245.jar
Binary file /usr/hdp/2.5.0.0-1245/hadoop/hadoop-common-2.7.3.2.5.0.0-1245.jar matches Also we verified size of this jar was less compared to the working one. RESOLUTION We copied this jar from the working host and datanode and nodemanager came up fine. We had no clue, from where this jar came, even of same version. But it was a good learning experience.
... View more
Labels:
01-04-2017
07:01 AM
Hi @Subodh Chettri, zeppelin server log will be helpful to identify the issue. If server is running and you are not able to open the web ui, might be some permission issue webapps folder. Please review the permission. /usr/hdp/<VERSION>/zeppelin/webapps/
... View more
01-03-2017
12:56 PM
thanks @krajguru for response. So it is expected behavior. Can we reduce this clock skew and is it advised to reduce it?
... View more
01-03-2017
12:21 PM
1 Kudo
I've modified the principal maxlife to 5min, as per the theory I should not be able to access the hdfs after 5min. But for me ticket is expiring after 10min. Any idea?
WARN ipc.Client: Exception encountered while connecting to the server :
javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]
modified maxlife using below command,
modprinc -maxlife "5minutes" hdfs-n@HWXBLR.COM
... View more
Labels:
- Labels:
-
Apache Hadoop
12-29-2016
12:06 PM
@Sandeep Nemuri, thanks for your reply. Does it has to be done on all the nodes, or from where I'm launching the spark jobs?
... View more
12-29-2016
11:59 AM
1 Kudo
Hi @Sandeep Nemuri, Can you please verify if spark.yarn.keytab and spark.yarn.principal is set in spark-thrift-sparkconf. This is a know bug. As STS is failing to renew the token once the cache expires, and there is no renew thread. Please add above properties and let me know if you are still seeing any issue.Thx
... View more
12-29-2016
11:43 AM
1 Kudo
As HDP comes with Python 2.6, but for spark jobs would like to use python 2.7 version. What all changes do we need to set to make only spark pick the installed 2.7 version. Thx
... View more
Labels:
- Labels:
-
Apache Spark
07-18-2016
08:00 AM
If you are using the java code, using hadoop class can delete the hdfs path
hdfs.delete(neworg.apache.hadoop.fs.Path(output),true) In spark you may try below, haven't tried myself though.
https://mail-archives.apache.org/mod_mbox/spark-user/201501.mbox/%3CCAHUQ+_ZwpDpfs1DaFW9zFFzJVW1PKTQ74kR2qbTqrBy7T96K9A@mail.gmail.com%3E
... View more