I just did a fresh clean install of CDH 5.11 and with Hive and Spark, everythin in dev mode - so embedded database, no HA, simple setup.
When I try to run spark-shell a got an exception:
Spark context available as sc (master = yarn-client, app id = application_1498152485620_0001).
17/06/22 19:30:59 WARN metastore.ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.1.0
17/06/22 19:30:59 WARN metastore.ObjectStore: Failed to get database default, returning NoSuchObjectException
SQL context available as sqlContext.
Is this some bug related to Spark 1.6?
I am sorry, but probably my mistake. I dont know why but now the sqlContext.sql("show tables").collect gathers the tables so I am able to access the metastore. The warning message is still there during spark-shell startup. But it works.