Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1980 | 07-09-2019 12:53 AM | |
| 11930 | 06-23-2019 08:37 PM | |
| 9186 | 06-18-2019 11:28 PM | |
| 10179 | 05-23-2019 08:46 PM | |
| 4603 | 05-20-2019 01:14 AM |
07-06-2016
11:58 PM
2 Kudos
Are you using hftp:// or webhdfs://? I'd recommend trying with the latter. For this specific exception in REST based copies, its usually not a fault with the network but a buggy state in the older Jetty used on the source cluster. Typically a rolling restart of the DataNodes will help resolve such a bad state of Jetty where it hangs up on a client mid-way during a response, causing the sudden EOF to the copying client in DistCp when it was expecting the rest of data.
... View more
07-06-2016
03:08 PM
The `delete` named column in your retailSKU table is actually a reserved word, and that's the central point of the issue. When Sqoop's building out the query, its currently not escaping the column names with backticks (``) which is necessary when not intending to mean using a reserved word ( https://dev.mysql.com/doc/refman/5.7/en/keywords.html). Would it be possible for you to alter the column name on the source retailSKU table, or pass a custom query instead which uses the right escape syntax, via --query: http://archive.cloudera.com/cdh5/cdh/5/sqoop/SqoopUserGuide.html#_free_form_query_imports ? Edit: Just noticed you've mentioned that free-form works just fine.
... View more
07-01-2016
04:16 AM
1 Kudo
While MultipleInputs was designed for such a thing, your requirement is unique in that you need to process the same input 2x but with different params each time. It seems a bit redundant to me given that you can do it in a single task run vs. 2x the I/O cost… But I believe the way you can solve your identifier problem is by writing your own InputFormat wrapper over the existing InputFormat, which generates special types of InputSplit objects (wrapper over regular FileSplit classes). These input splits need to add in your identifiers as an extra field, and you can extract and cast the same from your context.getInputSplit() in the map-end to then differentiate the input.
... View more
06-23-2016
06:42 PM
1 Kudo
When an RS exits, it logs a lot of content from the many threads it runs. Your log snippet is just the post-exit-decision messages, as you can note from the several "exiting" messages. Please instead pass us your entire log file, or look for the entire line and its following exception stack trace matching the word "FATAL", which would be in a line well before these logs appear.
... View more
06-20-2016
04:31 AM
2 Kudos
What directory are you looking under for the keytab files? Ensure to look under the most recent role-related directory. Note that you don't need to access these directories directly and that you can find the same process files under CM at CM -> [Service] -> Instances tab -> [Instance of the service roles you are interested in] -> Processes tab -> In the files section, keytab should be available for download.
... View more
06-20-2016
04:09 AM
The "^A" is a visual shell representation of the ASCII hexadecimal byte "01". You can run your command as so, making use of the actual byte representation than a visual helper: sqoop export --connect jdbc:oracle:thin:olap7964/olap7964@192.168.2.135:1521:ORCLOBIA11G --username XXXXXXX --password XXXXXXX --export-dir /user/hive/warehouse/abc.db/sach_sport --table SACH_SPORT --fields-terminated-by $'\x01' -m 1
... View more
06-14-2016
03:42 AM
1 Kudo
Thank you for double checking! The "-56" is the right layout version for DNs, so that's not the issue. There is instead a bug in that we'd left the sorted field out during conversion of the BlockReportContext objects: https://github.com/cloudera/hadoop-common/blob/cdh5.7.1-release/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/PBHelper.java#L2981-L2988 Internal report for this identified bug is CDH-40629, and the fix will be in the next update releases of 5.7.2 onwards and also in the 5.8.x series from 5.8.0 onwards. At present this shouldn't harm you, except the feature of HDFS-9260 won't really be in use and the perf. of NN would be the same as 5.6.x and lower.
... View more
06-13-2016
10:03 PM
Are these new logs that repeat over the day or is it an older entry? If you did a rolling upgrade, then it may be normal to see it during the period of upgrade, cause the sorting of BRs in the DN-side was a feature added only in 5.7.0+. Once the upgrade is complete and all DNs are restarted to use the newer code, you shouldn't observe the log entry anymore. Also check up your Live Nodes page on the NameNode Web UI to ensure all reporting DNs have their version story right.
... View more
06-13-2016
12:56 AM
2 Kudos
Ensure your Hive MetaStore service is up and running. If you use packages: service hive-metastore status If you use Cloudera Manager: CM -> Hive -> Instances -> Hive MetaStore Server page
... View more
06-12-2016
10:06 PM
2 Kudos
> Exact command cloudera gave is ADD JAR {{lib_dir}}/hive/lib/hive-contrib.jar; > 1)I am unable to find hive/lib directory on CDH 5 The {{lib_dir}} on CDH installed environments for Hive would either be /usr/lib/hive/ or /opt/cloudera/parcels/CDH/lib/hive/ (depending on packages or parcels being in use).
... View more