I'm trying to save data, which is processed by Spark, into HBase. Actually I'm using Phoenix to read and save my DataFrames:
// Read data from HBase table, e.g. in Zeppelin Notebook
val readDF = sqlContext.read
.format("org.apache.phoenix.spark")
.option("table", targetTable)
.option("zkUrl", zkUrl)
.load()
// Write DataFrame to HBase
myDF.write
.format("org.apache.phoenix.spark")
.mode("overwrite")
.option("table", targetTable)
.option("zkUrl", zkUrl)
.save()
Now I saw that there's also a Hortonworks Spark-HBase Connector:
The connector seems also to use Phoenix for SQL-like data access. My question now is, what are the differences between this connector and Phoenix? Why should I use the Spark-HBase Connector if this uses also Phoenix? Thank you!