Member since
07-30-2019
3404
Posts
1621
Kudos Received
1003
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 173 | 12-05-2025 08:25 AM | |
| 311 | 12-03-2025 10:21 AM | |
| 585 | 11-05-2025 11:01 AM | |
| 437 | 11-05-2025 08:01 AM | |
| 885 | 11-04-2025 10:16 AM |
06-05-2017
08:23 PM
@Joshua Adeleke HDF 2.1.4 (essentially HDF 2.1.3, plus the Controller service UI fix) will be out very very soon.
keep an eye out for it on the https://docs.hortonworks.com/ page. You can then just do an Ambari upgrade from HDF 2.1.3 to HDF 2.1.4. Thanks,
Matt
... View more
06-05-2017
02:05 PM
@Kiran Hebbar Hello, Your question is not very clear as to what you are looking for. I am going to assume you are asking how to view the metadata currently associated with a FlowFile passing through your NiFi dataflow(s). There are several ways to view this metadata: Right click on a connection that has queued data and select "List queue" from the context menu. Form the new UI that opens you will se a list of FlowFiles. Click on the icon to the left of any one of the FlowFiles to "view details" of that FlowFile. There you will find a "attributes" tab that list all the key/value pairs associated to this FlowFile. Use data provenance to perform a search on FlowFile events. "Data Provenance" can be found under the upper right corner hamburger menu in the NiFi UI. Click the search icon to open a "Search Events" UI where you can add criteria to limit the results (Provenance returns 1000 of the most recent events). From the final list use the same "view details" icon to the left of an event to open a new UI that will show the Attributes of your selected FlowFile. Use the LogAttribute processor. Add this processor anywhere in your dataflow. As FlowFile pass through this processor their FlowFile Attributes as they exist at the time of passing through this processor will be logged to the nifi-app.log. Keep in mind that this processor can greatly increase the size of your logs and require more space to store your logs. If you found this answer addressed your question, please mark it as accepted. Thanks, Matt
... View more
06-05-2017
12:40 PM
1 Kudo
@Paula DiTallo 1. Everything you configure in NiFi (Processors, connections, input ports, output ports, Remote Process groups, funnels, Controller services, reporting tasks, etc...) is contained within the flow.xml.gz file (by default located in NiFi's conf directory). You can clear the canvas in a couple ways: Select all components the canvas and click "delete" key. (Depending on connections, some components may not delete first time.) All connections must be absent of any queued data. All Processors must be stopped. This method will not delete any controller services, reporting tasks, or imported templates (these must be removed manually). Stop NiFi and delete the flow.xml.gz file. On next restart, a new blank flow.xml.gz file will be generated. Any FlowFiles that were still queued in NiFi will be deleted during start-up. This method will remove all components including controller services, reporting tasks, and imported templates. 2. User need to use the "upload template" icon found in the "Operate Panel" found to the left of the canvas. Once the template has been uploaded, it can be instantiated on to the canvas by dragging the "Template" icon form the top menu bar in the NIFi UI to the canvas. Thanks, Matt
... View more
06-02-2017
06:13 PM
3 Kudos
@Kiem Nguyen The documentation for the ListHDFS processor can be found imbedded in the imbedded documetation that is accessible via "help". Latest Apache version doc can also be found here: https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-hadoop-nar/1.2.0/org.apache.nifi.processors.hadoop.ListHDFS/index.html You need to get a copy of your core-site.xml and hdfs-site.xml files from your HDFS install and put them in the same place on every NiFi node to be referenced by this processor. Thanks, Matt
... View more
06-02-2017
12:21 PM
3 Kudos
@Joshua Adeleke Tested and confirmed a bug exists in HDF 2.1.3 that affects the Controller services UI. The fix for this bug will be in HDF 2.1.4. Unfortunately there is no easy workaround to this bug. You can still modify, delete, start, and stop controller services by interfacing directly with the NiFi Rest-API. Thanks, Matt
... View more
06-01-2017
03:23 PM
1 Kudo
@Alvaro Dominguez The primary node could change at anytime. You could use postHTTP and listenHTTP processor to route FlowFiles from multiple nodes to a single node. My concern would be heap usage to merge (zip) 160K FlowFiles on a single NiFi node. The FlowFile metadata for all those FlowFiles being zipped would be help in heap memory until the zip is complete. Any objection to having a zip of zips? In other words you could still create 4 unique zip files (1 per node each with unique filename), then send these zipped files to one node to be zipped once more in to a new zip with the single name you want written into HDFS. Thanks, Matt
... View more
06-01-2017
02:58 PM
1 Kudo
@Alvaro Dominguez The only way to create one single zip file is to have one node perform the zipping of all the files. This sounds less then ideal. How large are each of these individual zip files and how many FlowFiles on average go in to each zip file?
... View more
06-01-2017
02:40 PM
@Alvaro Dominguez Multiple nodes can write to the same path in HDFS, but not the same file at the same time. The lease error you saw above is moist likely the result of one node completing writing .example_1202.zip and then renaming it example_1202.zip. In between that time, a different node saw and then tried to start appending to .example_1202.zip, but it was moved/renamed before that could happen. It essentiall becomes a race condition since nodes do not communicate thsi kind of information with one another. You could write 4 zip files to HDFS every minute. You could just name each filename uniquely based on NiFi hostname writing file. Thanks,
Matt
... View more
06-01-2017
01:53 PM
1 Kudo
@Alvaro Dominguez Every node in a NiFi cluster runs its own copy of the cluster flow, has its own repositories, and works on its own set of FlowFiles. Nodes in a NiFi cluster are unaware of any FlowFiles being processed by other nodes in the cluster. What you are seeing is normal expected behavior of your dataflow. Thanks, Matt
... View more
06-01-2017
01:13 PM
@Oleksandr Solomko Any other ERROR or WARN log messages? --- Is this a standalone NiFi installation or a multi-node NiFi cluster? - If cluster, are all these FlowFiles queued on just one node? --- Is this Nifi secured (HTTPS or HTTP)? I can't reproduce locally. Thanks,
Matt
... View more