Member since
08-09-2017
9
Posts
0
Kudos Received
0
Solutions
09-26-2017
11:17 AM
09-07-2017
06:03 AM
The error is: Error in metadata: Cannot validate serde org.apache.hadoop.hive.contrib.serde2.JasonSerDe FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
... View more
- Tags:
- error
- Hadoop Core
- Hive
Labels:
- Labels:
-
Apache Hive
08-23-2017
12:25 PM
We are aware that all the data is stored in disk drives. The conundrum is that disk seeks times (latency of the disk operation) have not improved at the rate at which transfer rate (disk’s bandwidth) have reduced. So if a disk access operation comprises of more seeks, it would take a longer time to write or read through datasets than what it would take to stream through it. So for updating a small size of data to database, a traditional RDBMS will work just fine. However, in case of major data updates in database, MapReduce will be more optimized because it uses sort/merge to update complete database in one go. MapReduce trumps Traditional RDBMS on the following points: 1) MapReduce is a good fit for problems where there is a need to analyze a complete dataset in batch mode, particularly for ad hoc analysis. 2) Map Reduce suits application where data is written once and read many times. 3) Map Reduce is able to process Petabytes of data in a parallel fashion.
... View more