Member since
07-30-2019
944
Posts
197
Kudos Received
91
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1503 | 10-05-2021 01:53 PM | |
| 16313 | 09-23-2019 06:03 AM | |
| 6814 | 05-04-2019 08:42 PM | |
| 1502 | 06-11-2018 12:45 PM | |
| 12617 | 06-04-2018 01:11 PM |
05-03-2017
12:42 PM
1 Kudo
@Ravi Teja The flowfile isn't going out the FAILED connection because the session is rolled back to the incoming queue. I have highlighted that part of the log with the roll back information. Meaning the flowfile is penalized and then put back on the incoming queue for the PutSQL processor. If you want to write out the file that is causing the roll back, with a PutFile processor, there are ways to pull that flowfile out of the flow. You can use an RouteOnAttribute and put the uuid of the flowfile in as a property to get the flowfile and then route the flowfile out to the PutFile processor. For provenance events, there is a slight delay in the writing of events based on the configuration of NiFi. This is the property that controls how quickly the provenance events are available in the UI: nifi.provenance.repository.rollover.time -- The amount of time to wait before rolling over the latest data provenance information so that it is available in the User Interface. The default value is 30 secs. To see events for a particular processor, right click on the processor and a menu will pop up, select Data provenance Then another window will open displaying the provenance events for just that processor
... View more
05-02-2017
12:57 PM
@spdvnz There is no specific configuration for an AWS cloud instance. There are best practices when configuring NiFi, they can be found in the documentation embedded with NiFi and this HCC article is a good place to start also: Best Practices for high performance NiFi instance
... View more
05-01-2017
08:46 PM
2 Kudos
@Raj B The PutHDFS will "close" the file after each append. The writing of a file every minute will be less efficient than say writing a bigger file every ten minutes, but that is something you will have to determine which is more important to your use case. But even just waiting a minute is still more efficient then writing 100 files a sec.
... View more
05-01-2017
01:45 PM
@spdvnz NiFi will work "out of the box" in almost any environment as long as java is installed. Download the tar.gz file, extract the NiFi directories and then run the command bin/nifi.sh start from the NiFi root directory, and it will work. NiFi uses port 8080 for the UI by default. Is there something specific you want to know?
... View more
04-29-2017
12:19 AM
1 Kudo
@Bala Vignesh N V That is normal, now just click the "/nifi" link, it will take you to the UI.
... View more
04-28-2017
08:17 PM
1 Kudo
@Raj B I posted an answer to your other question before I saw this question. So, maybe a combination of the two methods would be the best approach. You could set a time limit on the merge of say 30 minutes or an hour, then check the file size in HDFS and if it isn't the desired size append, if it is then start writing to a new file in HDFS. That way you are still having a small impact on HDFS checking the file size only once every 30 to 60 minutes instead of every time a new file comes in.
... View more
04-28-2017
08:04 PM
2 Kudos
@Raj B It might be more efficient to just merge the incoming files to the desired size and then write the file to HDFS. The MergeContent processor can be configured to merge files to a desired size and then create one flow file from all of the merged files. For example: The MergeContent would be configured in the following way This would create a file between 1GB and 1.5GB. The processor can also be configured to use time limits on the merge file generation. But I think overall this approach would be less imposing on the HDFS system.
... View more
04-28-2017
05:38 PM
@Simon Jespersen Does the script have to be written using Python? There is a three part article written by @Matt Burgess that has some great examples: ExecuteScript Cookbook
... View more
04-27-2017
01:58 PM
1 Kudo
@Mohammed El Moumni
You can use the rest api to get the value of counter. You could create a script with the needed curl commands the get the value of the counter and then use the ExecuteScript processor to run the script. You can access the rest api document via your NiFi instance, the url will be similar to this http://nifi-host:port/nifi-docs/rest-api. For example, curl command to get counters in my flow, with output, is: curl 'http://nifi-host:port/nifi-api/counters' {"counters":{"aggregateSnapshot":{"generated":"15:12:55 BST","counters":[{"id":"46f421a7-d348-3e65-8a21-290cf1fc8fa1","context":"ExecuteScript (abde5770-015b-1000-0000-0000722fe055)","name":"my-counter","valueCount":0,"value":"0"},{"id":"bd62b47c-5124-35b1-926b-2149eb519f79","context":"ExecuteScript (abfdad0a-015b-1000-ffff-fffff0fcbeb9)","name":"my-counter","valueCount":56400,"value":"56,400"},{"id":"c6d58dcc-8171-3d63-adb2-8cc8e239cc9a","context":"Groovy script of counter (abfdad0a-015b-1000-ffff-fffff0fcbeb9)","name":"my-counter","valueCount":6930,"value":"6,930"},{"id":"776c5c43-40ee-372e-957c-f40b465d3326","context":"All ExecuteScript's","name":"my-counter"
... View more
04-27-2017
01:01 PM
@Mohammed El Moumni I owe you an apology. While trying to get a screenshot of the trash can, I discovered that in fact there is no way currently to delete the counter. I have opened a Apache Jira for this issue and here is the link: NIFI-3751 I've updated the post above with correct information.
... View more