Member since
07-19-2018
613
Posts
101
Kudos Received
117
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 5096 | 01-11-2021 05:54 AM | |
| 3422 | 01-11-2021 05:52 AM | |
| 8789 | 01-08-2021 05:23 AM | |
| 8385 | 01-04-2021 04:08 AM | |
| 36689 | 12-18-2020 05:42 AM |
03-30-2020
06:03 AM
@Former Member The solution is to use variables (in older versions (<= 1.9 and on) and parameters (1.10 and on).
... View more
03-30-2020
05:59 AM
@MPraveen Hello friend! I dont your sql schema but the error is wanting the real time version of the data (without the /). Glad you were able to solve it. Accept your own answer and get a solution credit!!
... View more
03-29-2020
06:05 AM
@mahfooz The mysql jar should be in the class path for the sqoop client. Also make sure the file ha the correct permissions.
... View more
03-29-2020
05:58 AM
@saaga119 The solution here is to create the external table in the format you need to query the raw xml data. Seems like you have this done already. Next create a native hive table. It could still be external too. The new table should be the schema for the fields you want with required ORC lines. You can also choose some other text format; for example: csv. Last INSERT INTO new_table SELECT * FROM external_table. The main idea here is to use external and staging tables, then INSERT INTO SELECT to fill the final table you want. Final table could then be optimized for performance (ORC). This kind of idea also allows multiple external/raw/staging tables able to be combined (select w/ join) into a single final optimized table (orc, compression, partitions, buckets, etc).
... View more
03-27-2020
06:16 AM
1 Kudo
@Gubbi It sounds like you have the getFile working and just having issues with the matching to target_file route? I would recommend to simply the match in RouteOnAttribute: ${filename:endsWith(${today})} If you are having problems with the GetFile, please share that proc configuration so we can see what you have going on there.
... View more
03-27-2020
06:01 AM
@Savi I think you only need one date format in your nifi expression language. Next, it seems you know what format it is upstream (source), and what it needs to be for downstream. Based on your expression need, you should play the flow up to the csv reader processor and then List Queue on the FlowFiles and inspect the attributes. Adjust the expression, and retest until the attribute format meets the required format. Inspecting the dates upstream won't show you what the date looks like after expression and before the reader. Often times I find myself testing a flow step by step, each step confirming the attributes needed are correct for the next step. Never moving forward to the next step until I am sure the attributes are correct after visual inspection.
... View more
03-27-2020
05:49 AM
@murali2425 It sure is possible. I believe anything is possible in NiFi!! The excel processor is: ConvertExcelToCSVProcessor I have also put an Excel to SQL template on my GitHub: https://github.com/steven-dfheinz/NiFi-Templates This template gets an excel file that contains product inventory data (sku, name, price, description, quantity, etc and routes the contents of the excel file to ConvertsExceltoCSV. Then it splits the csv file line by line, parses each line for quantity and sku and uses those attributes to build an insert query, which is finally executed via PutSQL. There are many different ways to deliver the Excel file too ConvertExceltoCsv, different ways to process the CSV, and different ways to insert into SQL. The above is just one way I have used in the past.
... View more
03-26-2020
07:50 AM
Update for Hue 4.x: I have finished initial required changes to for branch Hue.4.6.0 and successfully installed Hue 4.6.0 in HDP 3.1.4. There ended up being some big differences in the hue.ini for 4.x. Once these were resolved the only manual work needed for install in the node is creating the hue user and group and running the python to ignore_groupsusers_create. User Commands: useradd -g hue hue usermod -a -G wheel hue chown -R hue:hue /home/hue Reset/Remove Commands: rm -rf /var/lib/ambari-server/resources/stacks/HDP/3.1/services/HUE rm -rf /var/lib/ambari-agent/cache/stacks/HDP/3.1/services/HUE/ rm -rf /usr/local/hue rm -rf /usr/hdp/current/hue-server rm -rf /usr/local/hue-4.6.0/ Once installed it is possible to update/manage changes to hue.ini via ambari to enable HDFS, Hive, RDBMS, enable/disable Hue Plugins, or further configure Hue as necessary.
... View more
03-25-2020
05:50 AM
1 Kudo
@ParthiCyberPunk In HDP3 the only out of box option for hive is beeline. Additionally many third Party Tools can access hive via jdbc connections. I have been working on Hue with HDP3 and have had some success with Hue 3.x and HDP 3.x in production already. Here is article: https://community.cloudera.com/t5/Community-Articles/How-to-install-Hue-3-11-in-HDP-3-1/ta-p/291280 Here is the repo: https://github.com/steven-dfheinz/HDP3-Hue-Service I am currently workjng on hue 4.x and a management pack for HDP 2.x - 3.x
... View more
03-24-2020
08:44 AM
Repo Branch created for Hue 4.6.0. I am working on this right now but should have it operating here shortly. If you want Hue 4.6.0 be sure to download Branch Hue.4.6.0.
... View more