- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
CDH5 SerDe library 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe' is not supported.
- Labels:
-
Apache Hadoop
-
Apache Hive
-
Apache Impala
Created on ‎12-12-2014 05:25 PM - edited ‎09-16-2022 02:15 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi Cloudera People,
I am a beginner in Cloudera. I downloaded the VMQuickstart CDH5 and I try to do the tutorials but I can not finish the tutorial 3 because I have this error when I try to view in Impala the table " intermediate_access_logs"
Your query has the following error(s):
AnalysisException: Failed to load metadata for table: default.intermediate_access_logs CAUSED BY: TableLoadingException: Failed to load metadata for table: intermediate_access_logs CAUSED BY: InvalidStorageDescriptorException: Impala does not support tables of this type. REASON: SerDe library 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe' is not supported.
Thansk for helping
Created ‎12-13-2014 10:55 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
especially not in Impala. In that tutorial step you're actually using Hive
to do an ETL (extract transform load) job. The Apache logs are in a format
that is hard to query directly through SQL, so we use one of Hive's
extensions to express a regular expression to break up the fields more
explicitly. After this step, the intermediate table is not useful. It's the
second table you create (tokenized_access_logs) that should be queried from
Impala.
Created ‎12-13-2014 10:55 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
especially not in Impala. In that tutorial step you're actually using Hive
to do an ETL (extract transform load) job. The Apache logs are in a format
that is hard to query directly through SQL, so we use one of Hive's
extensions to express a regular expression to break up the fields more
explicitly. After this step, the intermediate table is not useful. It's the
second table you create (tokenized_access_logs) that should be queried from
Impala.
