Member since
10-01-2016
156
Posts
8
Kudos Received
6
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7962 | 04-04-2019 09:41 PM | |
3100 | 06-04-2018 08:34 AM | |
1437 | 05-23-2018 01:03 PM | |
2937 | 05-21-2018 07:12 AM | |
1798 | 05-08-2018 10:48 AM |
10-19-2019
02:45 AM
I deleted Sandbox and freshly imported it. It worked.
... View more
10-19-2019
01:14 AM
In HDP 2.6.5 I tried
[root@sandbox-hdp ~]# spark-shell --master yarn
But got following errors. It opens scala shell but cannot create spark session.
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
19/10/19 06:04:17 ERROR YarnClientSchedulerBackend: Yarn application has already exited with state FAILED!
19/10/19 06:04:17 ERROR TransportClient: Failed to send RPC 5080948039683175202 to /172.18.0.2:52542: java.nio.channels.ClosedChannelException
java.nio.channels.ClosedChannelException
at io.netty.channel.AbstractChannel$AbstractUnsafe.write(...)(Unknown Source)
19/10/19 06:04:17 ERROR YarnSchedulerBackend$YarnSchedulerEndpoint: Sending RequestExecutors(0,0,Map(),Set()) to AM was unsuccessful
java.io.IOException: Failed to send RPC 5080948039683175202 to /172.18.0.2:52542: java.nio.channels.ClosedChannelException
....
....
Caused by: java.nio.channels.ClosedChannelException
...
19/10/19 06:04:17 ERROR Utils: Uncaught exception in thread Yarn application state monitor
org.apache.spark.SparkException: Exception thrown in awaitResult:
...
Caused by: java.io.IOException: Failed to send RPC 5080948039683175202 to /172.18.0.2:52542: java.nio.channels.ClosedChannelException
...
19/10/19 06:04:17 ERROR SparkContext: Error initializing SparkContext.
java.lang.IllegalStateException: Spark context stopped while waiting for backend
...
console>:14: error: not found: value spark
import spark.implicits._
^
<console>:14: error: not found: value spark
import spark.sql
Then I looked at yarn logs:
19/10/19 06:21:35 INFO RMProxy: Connecting to ResourceManager at sandbox-hdp.hortonworks.com/172.18.0.2:8030
19/10/19 06:21:35 INFO YarnRMClient: Registering the ApplicationMaster
19/10/19 06:21:35 INFO YarnAllocator: Will request 2 executor container(s), each with 1 core(s) and 1408 MB memory (including 384 MB of overhead)
....
19/10/19 06:21:35 INFO RMProxy: Connecting to ResourceManager at sandbox-hdp.hortonworks.com/172.18.0.2:8030
19/10/19 06:21:35 INFO YarnRMClient: Registering the ApplicationMaster
19/10/19 06:21:35 INFO YarnAllocator: Will request 2 executor container(s), each with 1 core(s) and 1408 MB memory (including 384 MB of overhead)
... View more
Labels:
10-12-2019
10:49 PM
This (NULL issue) can be caused by data schema mismatch.
... View more
10-12-2019
10:10 PM
3 Kudos
Go to Ambari -> Hive -> Configs -> Advanced -> Custom hive-site Click Add Property Insert followings into the opening window: hive.security.authorization.sqlstd.confwhitelist.append=mapred.compress.map.output After saving, restart hive services. Then connect to beeline and set your parameter. I experienced a similar problem for mapreduce.job.reduces parameter and it worked.
... View more
10-08-2019
08:19 AM
My solution was a similar: in command line as a root systemctl stop mysqld
systemctl set-environment MYSQLD_OPTS="--skip-grant-tables --skip-networking"
systemctl start mysqld
mysql -u root in mysql shell use mysql;
update user set authentication_string=PASSWORD("hadoop") where User='root';
flush privileges;
quit After this change, I got trouble hive and ranger due to connection error to MySQL. Then I had to change hive and ranger mysql database root user password from Ambari. It is ok now.
... View more
10-04-2019
12:51 PM
Thank you @Shelton . I have just deleted the line specified in error message in /c/Users/user/.ssh/known_hosts It worked. I had used HDP 2.6.4 before this one. I think there was already a key for the same hostname, once delete it i was able to connect via ssh. This time added new key that belongs to HDP 2.6.5 Sandbox. In this case, Should I accept your solution? Because you offer additional solutions that I have not verified.
... View more
10-04-2019
07:13 AM
I have newly imported HDP 2.6.5 Sandbox after removing the 2.6.4, and tried to connect via ssh. My host machine Windows 10. My command:
ssh root@sandbox-hdp.hortonworks.com -p 2222
The error output:
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
@ WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! @
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
IT IS POSSIBLE THAT SOMEONE IS DOING SOMETHING NASTY!
Someone could be eavesdropping on you right now (man-in-the-middle attack)!
It is also possible that a host key has just been changed.
The fingerprint for the RSA key sent by the remote host is
SHA256:8F0VOvcJIA2FD3Ld90kv5FiD3270onl/+uDeDBNffQ0.
Please contact your system administrator.
Add correct host key in /c/Users/user/.ssh/known_hosts to get rid of this message.
Offending RSA key in /c/Users/user/.ssh/known_hosts:3
RSA host key for [sandbox-hdp.hortonworks.com]:2222 has changed and you have requested strict checking.
Host key verification failed.
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
07-31-2019
06:46 AM
Hi @Shu thank you. Adding spark-yarn_x.x.jar to maven repository solved the problem. But I have come across other errors. Anyway in here the problem was parsing the yarn and it is solved.
... View more
07-30-2019
09:34 AM
Hi @Shu I tried yarn-client and spark-yarn.jar But it can't pass the Could not parse Master URL: 'yarn' ERROR
... View more
07-26-2019
05:19 PM
I use IntelliJ IDE installed on my Windows 10 laptop and try to run spark job in yarn mode on my 5 node HDP 3.1.1 cluster. My codes: package p1
import org.apache.spark.sql.{SparkSession, functions => F}
import org.apache.log4j.{Logger, Level}
object SparkDeneme extends App {
Logger.getLogger("org").setLevel(Level.INFO)
val spark = SparkSession.builder()
.appName("SparkDeneme")
.master("yarn")
.config("spark.hadoop.fs.defaultFS","hdfs://node1.impektra.com:8020")
.config("spark.hadoop.yarn.resoursemanager.address","node1.impektra.com:8030")
.getOrCreate()
}
import spark.implicits._
val sc = spark.sparkContext
val dfFromList = sc.parallelize(List(1,2,3,4,5,6)).toDF("rakamlar")
// dfFromList.printSchema()
dfFromList.show() When I run get following error: 19/07/26 20:00:32 ERROR SparkContext: Error initializing SparkContext.
org.apache.spark.SparkException: Could not parse Master URL: 'yarn'
at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2744)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:492)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:933)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:924)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:924)
at p1.SparkDeneme$.delayedEndpoint$p1$SparkDeneme$1(SparkDeneme.scala:17)
at p1.SparkDeneme$delayedInit$body.apply(SparkDeneme.scala:8)
at scala.Function0$class.apply$mcV$sp(Function0.scala:34)
at scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.collection.immutable.List.foreach(List.scala:381)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35)
at scala.App$class.main(App.scala:76)
at p1.SparkDeneme$.main(SparkDeneme.scala:8)
at p1.SparkDeneme.main(SparkDeneme.scala) I tried to get help from this tutorial Anyone who has succeeded to run Spark YARN mode in IntelliJ?
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache YARN