Member since
01-04-2016
409
Posts
313
Kudos Received
35
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5699 | 01-16-2018 07:00 AM | |
1883 | 09-13-2017 06:17 PM | |
3743 | 09-13-2017 05:58 AM | |
2380 | 08-28-2017 07:16 AM | |
4153 | 05-11-2017 11:30 AM |
07-13-2016
11:55 AM
We are not using parallelism, could you please help from where I can reduce the cores. And yesterday this same code was working fine.
... View more
07-13-2016
11:24 AM
2 Kudos
I have 3 node cluster and trying to run the command. I am running follwoing command to run class file java -cp .:spark-assembly-1.6.1.2.4.2.0-258-hadoop2.7.1.2.4.2.0-258.jar:spark-csv_2.10-1.4.0.jar:commons-csv-1.1.jar SparkMainV4 "spark://xyz.abc.com:7077" "WD" "spark.executor.memory;6g,spark.shuffle.consolidateFile;false,spark.driver.memory;5g,spark.akka.frameSize;2047,spark.locality.wait;600,spark.network.timeout;600,spark.sql.shuffle.partitions;500" but getting error :- ERROR TaskSchedulerImpl: Lost executor 1 on xyz.abc.com: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages. Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: ResultStage 67 (saveAsTextFile at package.scala:179) has failed the maximum allowable number of times: 4. Most recent failure reason: org.apache.spark.shuffle.MetadataFetchFailedException: Missing an output location for shuffle 36
at org.apache.spark.MapOutputTracker$anonfun$org$apache$spark$MapOutputTracker$convertMapStatuses$2.apply(MapOutputTracker.scala:542)
at org.apache.spark.MapOutputTracker$anonfun$org$apache$spark$MapOutputTracker$convertMapStatuses$2.apply(MapOutputTracker.scala:538)
at scala.collection.TraversableLike$WithFilter$anonfun$foreach$1.apply(TraversableLike.scala:772)
at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771)
at org.apache.spark.MapOutputTracker$.org$apache$spark$MapOutputTracker$convertMapStatuses(MapOutputTracker.scala:538)
... View more
Labels:
- Labels:
-
Apache Spark
07-12-2016
08:39 AM
@chandramouli muthukumaran If issue is resolve, then lets close this.
... View more
06-17-2016
12:50 PM
@dnyanesh kulkarnni Yes right. Please install ntp, refer hte link http://www.openkb.info/2014/06/ntp-tips-for-hadoop-cluster.html
... View more
06-17-2016
09:45 AM
2 Kudos
@Sunile Manjee Steps to setup cluster:- There are some
pre-requisites required for hadoop that we need to take care of before we set
up the cluster. We will do them one by one now. 1) Setup
password-less ssh from master to slaves. 2) Update /etc/hosts
file 3) Update
it on each node to contain entries for all hosts in the cluster. 4) install Java 5) disable seliux and iptable 6)Download amabari repo wget -nv http://public-repo-1.hortonworks.com/ambari/centos6/2.x/updates/2.1.0/ambari.repo -O /etc/yum.repos.d/ambari.repo 7) install amabari yum install ambari-server 😎 run command to setup amabari ambari-server setup and press enter for all question 9) Now start the ambari server ambari-server start 10) Once service is started successfully then check UI at <ambari IP>:8080 11) You will see UI. Default login credentials are admin:admin (username:password) For the next steps follow the link:- http://docs.hortonworks.com/HDPDocuments/Ambari-2.1.0.0/bk_Installing_HDP_AMB/content/_launching_the_ambari_install_wizard.html If this is help you please accept the ans and lets close this.
... View more
06-17-2016
09:19 AM
4 Kudos
Please check this stpes:- 1. Make sure hostname is right, iptables, selinux are stop 2. Check ps -aef for each process and kill it from cli [if you are able to see the process running] 3. Start the process from ambari. let me know if it works.
... View more
06-15-2016
03:40 PM
@Marc Cardus If you you agree with answer, then lets close this one.
... View more
06-13-2016
02:43 PM
2 Kudos
I resolved the issue, with downgrade the version. steps what I perform stpe 1:- cat
/etc/yum.repos.d/HDP-TP.repo #VERSION_NUMBER=2.3.2.0-2950 [HDP-2.3.2.0] name=HDP Version -
HDP-2.3.2.0 baseurl=http://public-repo-1.hortonworks.com/HDP/centos7/2.x/updates/2.3.2.0 gpgcheck=1 gpgkey=http://public-repo-1.hortonworks.com/HDP/centos7/2.x/updates/2.3.2.0/RPM-GPG-KEY/RPM-GPG-KEY-Jenkins enabled=1 priority=1 [HDP-UTILS-1.1.0.20] name=HDP Utils
Version - HDP-UTILS-1.1.0.20 baseurl=http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.20/repos/centos7 gpgcheck=1 gpgkey=http://public-repo-1.hortonworks.com/HDP/centos7/2.x/updates/2.3.2.0/RPM-GPG-KEY/RPM-GPG-KEY-Jenkins enabled=1 priority=1 step2:- yum install
spark_2_3_2_0_2950* To check go to
/usr/hdp/2.3.2.0-2950/spark/sbin Step 3:- start
./start-all.sh
... View more