Member since
07-30-2019
3421
Posts
1624
Kudos Received
1010
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 50 | 01-13-2026 11:14 AM | |
| 181 | 01-09-2026 06:58 AM | |
| 504 | 12-17-2025 05:55 AM | |
| 565 | 12-15-2025 01:29 PM | |
| 561 | 12-15-2025 06:50 AM |
03-24-2017
02:10 PM
@mayki wogno
Is the same directory the same size of everyone of your zookeeper nodes? If not you may be having an issue on only one of your znodes. You should be able to shutdown the zookeeper node and purge all those files. The pertain files will be re-written from the other znodes in the zookeeper cluster when it rejoins the zookeeper cluster. Zookeeper is storing information about who is your current cluster coordinator, primary node, and any cluster wide state various from various processor in your dataflows. I am assuming you are running the embedded zookeeper here. In that case the zookeeper.properties file should control the auto purge of the snapshots through the following properties: autopurge.purgeInterval=24
autopurge.snapRetainCount=30 The transaction logs should be handle via routine maintenance which you can find here: http://archive.cloudera.com/cdh4/cdh/4/zookeeper/zookeeperAdmin.html#sc_maintenance Thanks, Matt
... View more
03-24-2017
12:33 PM
@Praveen Singh Standard out from your script is written to the content of the FlowFile generated by the ExecuteProcess NiFi processor. So perhaps just tweaking your script to write to standard out rather then a file on disk is all you need to do.
... View more
03-23-2017
02:48 PM
@Sanaz Janbakhsh If you found the information provided useful, please accept that answer. Thank you, Matt
... View more
03-23-2017
02:44 PM
@Sanaz Janbakhsh It is "zero master clustering". All nodes in an HDF 2.0 (NiFi 1.x) cluster run the dataflow and do work on FlowFiles. An election is conducted and at completion of that election one node will be elected as the cluster coordinator and one node will be elected as the primary node (run primary node only configured processors). Which node in the cluster is assigned these roles can change at anytime should the previously elected node should stop sending heartbeats in the configured threshold. It also possible for same node to be elected both roles. This also means that any node in a HDF 2.0 cluster can be used for establishing Site-to-Site (S2S) connections. Ind old NiFi S2S to a cluster required that the RPG point at the NCM. Thanks, Matt
... View more
03-23-2017
02:12 PM
2 Kudos
@Diego Labrador Anytime you encounter the message "Unable to perform the desired action due to insufficient permissions. Contact the system administrator.", you are having an authorization issue. Authentication issues present different errors. You should inspect your nifi-user.log while trying to access the UI to see what the exact string is being passed to the authorizer. By default with ldap as your configured login identity provider, the full DN for the user who logged in is passed to the authorizer. By the looks of the above you configured on the CN= as your initial admin identity. The string passed to the authorizer will be shown in nifi-user.log and must matcha exactly (Case sensitive and spaces count as valid characters also. Thanks, Mat
... View more
03-23-2017
02:02 PM
@Sanaz Janbakhsh HDF 2.0 is based off Apache NiFi 1.0 which no longer has an NCM (NCM based clusters only exists in HDF 1.x or Apache NiFi 0.x versions). HDF 2.0 is a zero master cluster which requires Zookeeper (min 3 ZK nodes for Quorum) for cluster coordinator and primary node designations and for storing your cluster wide state. - Is physical servers recommended for HDF other than VM? I do recommend physical servers over VM for NiFi. Depending on the dataflow(s) you design (which processor and controller service components you use), The load put on your server can go form very light to very heavy. - How many minimum nodes needs to be deployed for the clustering? There is no minimum number of hosts in a NiFi cluster. You can actually even stand up a 1 node cluster (Pointless and actual will perform poorer the a standalone NiFi because of additional cluster overhead that is added). I suggest starting with a 3 node cluster to spread out your load and provide coverage if a node is lost. You can add additional node to an existing Nifi cluster later with minimal effort. - What is the minimum hardware requirements per each node? Not knowing exactly what you plan on doing in your dataflow with regards to your 35,000 FlowFiles per minute, it is difficult to make any CPU suggestions. Generally speaking it is good practice to setup a POC and see how it scales. The fact that you are working with a large number of very small files, NiFi JVM heap usage could potentially be high, so making sure you have enough memory on each node to give NiFi at least 8 GB of heap to start with. You will need additional memory for the OS and any other services running on these host other then NiFi. Thanks, Matt
... View more
03-23-2017
01:44 PM
@Praveen Singh Try using the "ExecuteProcess" processor. Property Value Command ssh Command Arguments -i "<path to private key>" <user>@<remotehost> python <script> &
Thanks, Matt
... View more
03-23-2017
01:05 PM
1 Kudo
@Simon Jespersen The "Host Key File" property is used to specify the file containing your trusted hosts (commonly named known_hosts and found by default in the .ssh directory). It is not the key you are using to connect with. This property works in conjunction with the "Strict Host Key Checking" property when it is set to "true" You are getting a key does not exist because this property ins un related to "Private Key Path" so NiFi is looking in its base default directory for this file. I believe what you are trying to do has nothing to do with host key checking. You want to configure the "Private Key Path" as the full path to your private key including the key name itself. Thank you, Matt
... View more
03-23-2017
12:39 PM
3 Kudos
@Praveen Singh NiFi offers a variety of processors that support executing commands or scripts locally to the NiFi installation. In order to execute a script remotely to NiFi you would still need to have a local script that would be called by the processor that could then go perform that task. Thank you, Matt
... View more
03-23-2017
12:35 PM
@Roger Young I assume the Remote Process Group (RPG) is on MiNiFi. Try checking the following things: 1. NiFi will not fail startup if it cannot bind to the configured nifi.renote.input.socket.port. Make sure there are no other applications or NiFi instances running on this box that way have already bound to this port. Try using a different port and restarting 2. Since you are just testing, try leaving nifi.remote.input.host blank or setting it to 127.0.0.1 instead on your Windows 10 install. Thanks, Matt
... View more