Member since
07-30-2019
3427
Posts
1632
Kudos Received
1011
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 69 | 01-27-2026 12:46 PM | |
| 482 | 01-13-2026 11:14 AM | |
| 983 | 01-09-2026 06:58 AM | |
| 899 | 12-17-2025 05:55 AM | |
| 960 | 12-15-2025 01:29 PM |
06-07-2017
02:29 PM
1 Kudo
@J. D. Bacolod Have you considered using the PutDistributedMapCache and GetDistributedMapCache processors? Have two separate dataflows. One runs on a cron and is responsible for obtaining the token and write that token to the distirbutedMapCache using the putDistirbutedMapCache processor. The Second flow is for doing all your other operations using that token. Just before the invokeHTTP processor add a GetDistibutedMapCache processor that reads the token from the distributed map cache in to a FlowFile attribute. You then use that attribute to pass the token in your connections. One thing to keep in mind is that it is possible that a new token may be retrieved after a FlowFile had already retrieved the old token from the distirbutedMapCache. This would result in auth failure. So you will want your flow to loop back to GetDistributedMapChace processor to get latest key on auth failure on your invokeHTTP processor. This flow does not keep track in any way when a token expires, but if you know how long a token is good for you can set your cron accordingly. Thanks, Matt
... View more
06-07-2017
01:03 PM
2 Kudos
@Anthony Murphy Is this a NiFi cluster or standalone NIFi instance? Make sure that the following property is set to "true" in the nifi.properties file on every instance of NiFi: nifi.flowcontroller.autoResumeState=true If this property is set to false it will trigger all components to come up stopped when Nifi is restarted. Thanks, Matt
... View more
06-06-2017
09:02 PM
@Alvin Jin My suggestion would be use something like: https://regex101.com/ You can enter your regex and sample test you want to run it against. Matt
... View more
06-05-2017
08:23 PM
@Joshua Adeleke HDF 2.1.4 (essentially HDF 2.1.3, plus the Controller service UI fix) will be out very very soon.
keep an eye out for it on the https://docs.hortonworks.com/ page. You can then just do an Ambari upgrade from HDF 2.1.3 to HDF 2.1.4. Thanks,
Matt
... View more
06-05-2017
02:05 PM
@Kiran Hebbar Hello, Your question is not very clear as to what you are looking for. I am going to assume you are asking how to view the metadata currently associated with a FlowFile passing through your NiFi dataflow(s). There are several ways to view this metadata: Right click on a connection that has queued data and select "List queue" from the context menu. Form the new UI that opens you will se a list of FlowFiles. Click on the icon to the left of any one of the FlowFiles to "view details" of that FlowFile. There you will find a "attributes" tab that list all the key/value pairs associated to this FlowFile. Use data provenance to perform a search on FlowFile events. "Data Provenance" can be found under the upper right corner hamburger menu in the NiFi UI. Click the search icon to open a "Search Events" UI where you can add criteria to limit the results (Provenance returns 1000 of the most recent events). From the final list use the same "view details" icon to the left of an event to open a new UI that will show the Attributes of your selected FlowFile. Use the LogAttribute processor. Add this processor anywhere in your dataflow. As FlowFile pass through this processor their FlowFile Attributes as they exist at the time of passing through this processor will be logged to the nifi-app.log. Keep in mind that this processor can greatly increase the size of your logs and require more space to store your logs. If you found this answer addressed your question, please mark it as accepted. Thanks, Matt
... View more
06-05-2017
12:40 PM
1 Kudo
@Paula DiTallo 1. Everything you configure in NiFi (Processors, connections, input ports, output ports, Remote Process groups, funnels, Controller services, reporting tasks, etc...) is contained within the flow.xml.gz file (by default located in NiFi's conf directory). You can clear the canvas in a couple ways: Select all components the canvas and click "delete" key. (Depending on connections, some components may not delete first time.) All connections must be absent of any queued data. All Processors must be stopped. This method will not delete any controller services, reporting tasks, or imported templates (these must be removed manually). Stop NiFi and delete the flow.xml.gz file. On next restart, a new blank flow.xml.gz file will be generated. Any FlowFiles that were still queued in NiFi will be deleted during start-up. This method will remove all components including controller services, reporting tasks, and imported templates. 2. User need to use the "upload template" icon found in the "Operate Panel" found to the left of the canvas. Once the template has been uploaded, it can be instantiated on to the canvas by dragging the "Template" icon form the top menu bar in the NIFi UI to the canvas. Thanks, Matt
... View more
06-02-2017
06:13 PM
3 Kudos
@Kiem Nguyen The documentation for the ListHDFS processor can be found imbedded in the imbedded documetation that is accessible via "help". Latest Apache version doc can also be found here: https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.2.0/org.apache.nifi.processors.hadoop.ListHDFS/index.html You need to get a copy of your core-site.xml and hdfs-site.xml files from your HDFS install and put them in the same place on every NiFi node to be referenced by this processor. Thanks, Matt
... View more
06-02-2017
12:21 PM
3 Kudos
@Joshua Adeleke Tested and confirmed a bug exists in HDF 2.1.3 that affects the Controller services UI. The fix for this bug will be in HDF 2.1.4. Unfortunately there is no easy workaround to this bug. You can still modify, delete, start, and stop controller services by interfacing directly with the NiFi Rest-API. Thanks, Matt
... View more
06-01-2017
03:23 PM
1 Kudo
@Alvaro Dominguez The primary node could change at anytime. You could use postHTTP and listenHTTP processor to route FlowFiles from multiple nodes to a single node. My concern would be heap usage to merge (zip) 160K FlowFiles on a single NiFi node. The FlowFile metadata for all those FlowFiles being zipped would be help in heap memory until the zip is complete. Any objection to having a zip of zips? In other words you could still create 4 unique zip files (1 per node each with unique filename), then send these zipped files to one node to be zipped once more in to a new zip with the single name you want written into HDFS. Thanks, Matt
... View more
06-01-2017
02:58 PM
1 Kudo
@Alvaro Dominguez The only way to create one single zip file is to have one node perform the zipping of all the files. This sounds less then ideal. How large are each of these individual zip files and how many FlowFiles on average go in to each zip file?
... View more