- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Spark job reeturns empty rows from HBase
- Labels:
-
Apache HBase
-
Apache Spark
Created ‎07-03-2018 03:17 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi Community,
I'm running a basic spark job which reads from an HBase table.
I can see the job is getting complete without any error, but in output I get the empty rows.
Will appreciate any help.
Below is my code
object objectName { def catalog = s"""{ |"table":{"namespace":"namespaceName", "name":"tableName"}, |"rowkey":"rowKeyAttribute", |"columns":{ |"Key":{"cf":"rowkey", "col":"rowKeyAttribute", "type":"string"}, |"col1":{"cf":"cfName", "col":"col1", "type":"bigint"}, |"col2":{"cf":"cfName", "col":"col2", "type":"string"} |} |}""".stripMargin def main(args: Array[String]) { val spark = SparkSession.builder() .appName("dummyApplication") .getOrCreate() val sc = spark.sparkContext val sqlContext = spark.sqlContext import sqlContext.implicits._ def withCatalog(cat: String): DataFrame = { sqlContext .read .options(Map(HBaseTableCatalog.tableCatalog -> cat)) .format("org.apache.spark.sql.execution.datasources.hbase") .load() } }
Created ‎07-03-2018 05:35 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Did you check out the docs?
Did you look at this other HCC post on a similar topic:
https://community.hortonworks.com/questions/49743/read-hbase-table-by-using-sparkscala.html
Created ‎07-03-2018 05:35 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Did you check out the docs?
Did you look at this other HCC post on a similar topic:
https://community.hortonworks.com/questions/49743/read-hbase-table-by-using-sparkscala.html

- « Previous
-
- 1
- 2
- Next »