Member since
07-19-2018
613
Posts
101
Kudos Received
117
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 5687 | 01-11-2021 05:54 AM | |
| 3812 | 01-11-2021 05:52 AM | |
| 9487 | 01-08-2021 05:23 AM | |
| 9288 | 01-04-2021 04:08 AM | |
| 38601 | 12-18-2020 05:42 AM |
04-02-2020
11:45 AM
@Gubbi I think your ListFile proc is still executing 0 sec. Reference our private message.
... View more
04-02-2020
06:32 AM
1 Kudo
@Gubbi The next solution here is to just add each route for today, yesterday, day before yesterday. Then route all 3 to the next proc. Anything not matching won't be routed.
... View more
04-02-2020
06:29 AM
1 Kudo
@Gubbi The solution here is a now minute 24 hours: Yesterday: ${now():minus(86400000):format('MM-dd-yyyy hh:mm:ss') } Day Before Yesterday: ${now():minus(86400000):minus(86400000):format('MM-dd-yyyy hh:mm:ss') }
... View more
03-30-2020
07:51 PM
Thanks Steven
... View more
03-29-2020
05:58 AM
@saaga119 The solution here is to create the external table in the format you need to query the raw xml data. Seems like you have this done already. Next create a native hive table. It could still be external too. The new table should be the schema for the fields you want with required ORC lines. You can also choose some other text format; for example: csv. Last INSERT INTO new_table SELECT * FROM external_table. The main idea here is to use external and staging tables, then INSERT INTO SELECT to fill the final table you want. Final table could then be optimized for performance (ORC). This kind of idea also allows multiple external/raw/staging tables able to be combined (select w/ join) into a single final optimized table (orc, compression, partitions, buckets, etc).
... View more
03-28-2020
09:26 AM
1 Kudo
great and Thanks it was very useful for my work. Here I will be having data in the excel, I need to extract the column names(1st row) and form the create the table with primary key and foreign key and data(2nd row) in the database(mysql) from nifi.
... View more
03-27-2020
06:01 AM
@Savi I think you only need one date format in your nifi expression language. Next, it seems you know what format it is upstream (source), and what it needs to be for downstream. Based on your expression need, you should play the flow up to the csv reader processor and then List Queue on the FlowFiles and inspect the attributes. Adjust the expression, and retest until the attribute format meets the required format. Inspecting the dates upstream won't show you what the date looks like after expression and before the reader. Often times I find myself testing a flow step by step, each step confirming the attributes needed are correct for the next step. Never moving forward to the next step until I am sure the attributes are correct after visual inspection.
... View more
03-25-2020
06:48 PM
Thanks @stevenmatison . worked perfectly..
... View more
03-25-2020
12:27 PM
@MattWho Is HDF 3.5 already released? If not, do you know when it is planned to be released? I saw page with release notes, but repository locations are still not updated. https://docs.cloudera.com/HDPDocuments/HDF3/HDF-3.5.0/release-notes/content/hdf_repository_locations.html
... View more
03-25-2020
11:29 AM
Hive view is no longer available in Ambari 2.7.x version (required for HDP 3). And is deprecated in support of DAS/DAS lite. Alternatively you can use the JDBC tools like DBvisulaizer or Squirrel or Hue.
... View more