Member since
04-26-2016
12
Posts
3
Kudos Received
0
Solutions
08-03-2016
09:17 PM
Hi We are thinking on having dedicated edge per project for our data lake. Each project will have a vm on which we install the required clients. Anyone is doing this ? Any problems or issues that we should be aware of with this configuration ?
... View more
07-27-2016
09:02 PM
1 Kudo
Hi, Does anyone know a good GUI for HBase for creating and querying tables ? something like Hue HBase browser for HDP ? any plan for HBase Ambari view ? Thanks
... View more
Labels:
- Labels:
-
Apache HBase
06-24-2016
05:44 AM
@sujitha sanku Thanks
I am talking about libraries that doesn't come with Spark by default like spark-csv. This code works with Spark-shell but not with Zeppelin (same thing if I use Pyspark): import org.apache.spark.sql.SQLContext
val df = sqlContext.read.format("com.databricks.spark.csv").option("header", "true").option("inferSchema", "true").load("/tmp/sales.csv")
df.printSchema()
val selectedData = df.select("customerId", "itemId")
selectedData.collect() Should I add import statement ? why this is working in Spark directly
... View more
06-23-2016
10:14 PM
I want to add a library and use it in Zeppelin (ex. Spark-csv). I succeeded in adding it to Spark and using it by putting my Jar in all nodes and adding spark.jars='path-to-jar' in conf/spark-defaults.conf. However when I call the library from Zeppelin it doesn't work (class not found). From my understanding Zeppelin do a Spark-submit so if the package is already added in Spark it should work. Also, I tried adding using export
SPARK_SUBMIT_OPTIONS=”--jars
/path/mylib1.jar,/path/mylib2.jar" to zeppelin-env.sh but same problem. Has anyone suceeded in adding libraries to Zeppelin ? have you seen this problem ?
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache Zeppelin