Member since
01-19-2017
5
Posts
1
Kudos Received
0
Solutions
09-14-2016
03:36 PM
Tested below in AWS. Looks good. Thank you //read error JSON file from cluster 1 val erorDF = spark.read.json("hdfs://master:8020/user/ubuntu/error.json") erorDF.registerTempTable("erorDFTBL") //read file from cluster 2 val erorDF2 = spark.read.json("hdfs://master2:8020/user/ubuntu/errors") erorDF2.registerTempTable("erorDFTBL2")
... View more
09-14-2016
02:38 PM
Thank you Becker. Will there be any setup I need to do in Zeppelin. I am running my Zeppelin in cluster 1.
... View more
09-13-2016
07:46 PM
1 Kudo
Hi I want to source data from two hadoop clusters and join in Spark. Will it possible as shown below //data from cluster1 val erorDF = spark.read.json("hdfs://master:8020//user/ubuntu/error.json") erorDF.registerTempTable("erorDFTBL") //data from cluster2 val erorDF2 = spark.read.json("hdfs://master2:8020//user/ubuntu/error.json") erorDF2.registerTempTable("erorDFTBL2")
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Spark
07-14-2016
06:15 PM
Hi Spark 2.0 now had structured streaming. How it is different from NIFI file streaming.
... View more
Labels:
- Labels:
-
Apache Spark