Member since
10-01-2016
156
Posts
8
Kudos Received
6
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 11843 | 04-04-2019 09:41 PM |
03-25-2020
06:28 AM
Thank you very much. This is the one that satisfies me. Documents are expected to make clear and simple things, not complicated.
... View more
03-23-2020
01:39 AM
I know thank you but I still don't understand why we add same property (dfs.datanode.balance.max.concurrent.moves) on a different section DataNode Advanced Configuration Snippet (Safety Valve) for hdfs-site.xml Balancer Advanced Configuration Snippet (Safety Valve) for hdfs-site.xml although same property (dfs.datanode.balance.max.concurrent.moves) already exists in Cloudera Manager. Is CM supposed to refuse this addition?
... View more
03-22-2020
12:15 PM
Yes, you are right. I have not realized that. But if dfs.datanode.ec.reconstruction.xmits.weight is already in hdfs-site.xml why Cloudera document makes us add the same property for balancer and DataNode again, what is the point?
... View more
03-22-2020
02:49 AM
I am trying to rebalance hdfs with Cloudera Manager 6.3 with HDFS Balancer Document
It says add the same property dfs.datanode.balance.max.concurrent.moves into different section
DataNode Advanced Configuration Snippet (Safety Valve) for hdfs-site.xml
Balancer Advanced Configuration Snippet (Safety Valve) for hdfs-site.xml
But before adding the property I searched and saw dfs.datanode.balance.max.concurrent.moves was already there. Nevertheless, I did what the document says. After adding properties Cloudera Manager asked me to restart/redeploy stale configurations. Before restart, I saw totally different properties added.
I don't understand although we seem to add the same property why different properties are added to hdfs-site.xml?
... View more
Labels:
- Labels:
-
Cloudera Manager
-
HDFS
02-28-2020
05:06 AM
In CDH 6.X I can't find Advanced spark2-metrics-properties in Spark config. Should I create manually?
... View more
10-04-2019
12:51 PM
Thank you @Shelton . I have just deleted the line specified in error message in /c/Users/user/.ssh/known_hosts It worked. I had used HDP 2.6.4 before this one. I think there was already a key for the same hostname, once delete it i was able to connect via ssh. This time added new key that belongs to HDP 2.6.5 Sandbox. In this case, Should I accept your solution? Because you offer additional solutions that I have not verified.
... View more
10-04-2019
07:13 AM
I have newly imported HDP 2.6.5 Sandbox after removing the 2.6.4, and tried to connect via ssh. My host machine Windows 10. My command:
ssh root@sandbox-hdp.hortonworks.com -p 2222
The error output:
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
@ WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! @
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
IT IS POSSIBLE THAT SOMEONE IS DOING SOMETHING NASTY!
Someone could be eavesdropping on you right now (man-in-the-middle attack)!
It is also possible that a host key has just been changed.
The fingerprint for the RSA key sent by the remote host is
SHA256:8F0VOvcJIA2FD3Ld90kv5FiD3270onl/+uDeDBNffQ0.
Please contact your system administrator.
Add correct host key in /c/Users/user/.ssh/known_hosts to get rid of this message.
Offending RSA key in /c/Users/user/.ssh/known_hosts:3
RSA host key for [sandbox-hdp.hortonworks.com]:2222 has changed and you have requested strict checking.
Host key verification failed.
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
07-31-2019
06:46 AM
Hi @Shu thank you. Adding spark-yarn_x.x.jar to maven repository solved the problem. But I have come across other errors. Anyway in here the problem was parsing the yarn and it is solved.
... View more
07-30-2019
09:34 AM
Hi @Shu I tried yarn-client and spark-yarn.jar But it can't pass the Could not parse Master URL: 'yarn' ERROR
... View more
07-26-2019
05:19 PM
I use IntelliJ IDE installed on my Windows 10 laptop and try to run spark job in yarn mode on my 5 node HDP 3.1.1 cluster. My codes: package p1
import org.apache.spark.sql.{SparkSession, functions => F}
import org.apache.log4j.{Logger, Level}
object SparkDeneme extends App {
Logger.getLogger("org").setLevel(Level.INFO)
val spark = SparkSession.builder()
.appName("SparkDeneme")
.master("yarn")
.config("spark.hadoop.fs.defaultFS","hdfs://node1.impektra.com:8020")
.config("spark.hadoop.yarn.resoursemanager.address","node1.impektra.com:8030")
.getOrCreate()
}
import spark.implicits._
val sc = spark.sparkContext
val dfFromList = sc.parallelize(List(1,2,3,4,5,6)).toDF("rakamlar")
// dfFromList.printSchema()
dfFromList.show() When I run get following error: 19/07/26 20:00:32 ERROR SparkContext: Error initializing SparkContext.
org.apache.spark.SparkException: Could not parse Master URL: 'yarn'
at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2744)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:492)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:933)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:924)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:924)
at p1.SparkDeneme$.delayedEndpoint$p1$SparkDeneme$1(SparkDeneme.scala:17)
at p1.SparkDeneme$delayedInit$body.apply(SparkDeneme.scala:8)
at scala.Function0$class.apply$mcV$sp(Function0.scala:34)
at scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.collection.immutable.List.foreach(List.scala:381)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35)
at scala.App$class.main(App.scala:76)
at p1.SparkDeneme$.main(SparkDeneme.scala:8)
at p1.SparkDeneme.main(SparkDeneme.scala) I tried to get help from this tutorial Anyone who has succeeded to run Spark YARN mode in IntelliJ?
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache YARN