Member since
05-11-2016
29
Posts
1
Kudos Received
3
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7369 | 09-25-2017 02:06 PM | |
12526 | 08-16-2016 01:41 PM | |
3442 | 05-18-2016 06:30 AM |
08-17-2018
11:11 AM
I have this same problem, I too tried follow the method of @Vinod369 and not resolved
... View more
01-16-2018
10:01 PM
where should we add this warn??
... View more
09-25-2017
02:06 PM
Sorry, could not focus on this, was busy with production activities. Finally, I could able to run it successfully with below configurations, Jar / py name : ${nameNode}/user/solution.jar Main Class : Module.final_solution Options List : --conf spark.yarn.jar=local:/opt/cloudera/parcels/CDH/lib/spark/lib/spark-assembly.jar Properties: Spark Master : yarn Mode : cluster App name : Final Solution
... View more
08-16-2017
03:07 AM
Great Suggestion! Thanks for that! But i would like to ask one question that, if i want to have a struct or array fields in target, then how i should transform the mysql data, so that it will fit in HCatalog schema. The need here is to just have nested data from other collection instead of foreign key representation. Currently we are using sqoop import only, and we are trying to modify the query so that it will be accepted by hcat schema. Thanks & Regards, Mahendra
... View more
07-28-2016
06:15 PM
The Cloudera ODBC connector is available for Windows, and .NET does support ODBC: http://www.cloudera.com/downloads/connectors/hive/odbc.html There's no direct client for HDFS in .NET, but HDFS offers a REST API via its WebHDFS component. This REST API is documented at http://archive.cloudera.com/cdh5/cdh/5/hadoop/hadoop-project-dist/hadoop-hdfs/WebHDFS.html, and you can use your .NET HTTP client to make use of it.
... View more
06-24-2016
08:34 AM
Your understanding is correct. You either need to ensure flume can write to that directory, or create a directory that flume owns and can write to. -pd
... View more
06-07-2016
04:23 PM
The flume http source is for creating a REST API that you can post data to from the upstream sender. If you are looking for a source that will consume from SQL server via accessing the SQL server API, you'll need to write a custom source for that, or possibly try this: https://github.com/keedio/flume-ng-sql-source Additionally, if you don't need real time processing, you may want to consider using sqoop to import data via batch processing, and it can handle incremental updates. -pd
... View more
05-26-2016
05:15 AM
Thank you for detail reply. I have initiated Flume as service on Edge node and its as expected.
... View more
05-25-2016
10:38 AM
This documentation goes over stopping and starting flume when not using Cloudera Manager. This assumes you are running packages and not parcels on this edge node: http://www.cloudera.com/documentation/enterprise/latest/topics/cdh_ig_flume_run.html
... View more