- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Spark failed to get database default
- Labels:
-
Apache Hive
-
Apache Spark
Created on 06-22-2017 10:35 AM - edited 09-16-2022 04:48 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi,
I just did a fresh clean install of CDH 5.11 and with Hive and Spark, everythin in dev mode - so embedded database, no HA, simple setup.
When I try to run spark-shell a got an exception:
Spark context available as sc (master = yarn-client, app id = application_1498152485620_0001).
17/06/22 19:30:59 WARN metastore.ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.1.0
17/06/22 19:30:59 WARN metastore.ObjectStore: Failed to get database default, returning NoSuchObjectException
SQL context available as sqlContext.
Is this some bug related to Spark 1.6?
Thanks
Tomas
Created 06-22-2017 10:03 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@mbigelowi agree with you on this.
Could you check your Metastore / HS2 status
sudo service hive-metastore status
sudo service hive-server2 status
Created 06-22-2017 11:01 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created 06-22-2017 06:46 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
The database it is trying to access is the backend to the Hive Metastore. Are you able to access and view databases and tables in Hive?
Created 06-23-2017 02:59 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I am sorry, but probably my mistake. I dont know why but now the sqlContext.sql("show tables").collect gathers the tables so I am able to access the metastore. The warning message is still there during spark-shell startup. But it works.
Created 06-22-2017 10:03 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@mbigelowi agree with you on this.
Could you check your Metastore / HS2 status
sudo service hive-metastore status
sudo service hive-server2 status
Created 06-03-2019 07:48 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Still not clear on what the solution is for this error. We are getting the same error. Please help !!
Created 06-05-2019 12:15 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Could you let me know what is the issue you are facing ? whats the error ?
