Member since
07-30-2019
3467
Posts
1641
Kudos Received
1016
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 91 | 05-04-2026 05:20 AM | |
| 426 | 03-23-2026 05:44 AM | |
| 325 | 02-18-2026 09:59 AM | |
| 571 | 01-27-2026 12:46 PM | |
| 1002 | 01-20-2026 05:42 AM |
03-19-2026
05:37 AM
@Vishesh Sorry, but there is not enough information to give a definitive cause for your issue. There is typically full stack traces to go with these types of logged exceptions which can tell us more, but even then it may not be enough information still. As i mentioned before there are numerous improvements between Apache NIFi 1.23 and the last Apache NIFi 1.x release version 1.28. I'd strongly encourage you to upgrade to see if your issue persists. You may be hitting this bug NIFI-12232 which was addressed in versions 1.26+. Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
03-17-2026
06:29 AM
1 Kudo
@Vishesh It is not possible to say exactly what issue you may have encountered here. Do you still have the complete stack trace that followed the node disconnection exception? It is likely to have some "Caused by:.." lines in the full stack trace that may help. Any changes being done when the disconnection occurred? Stopping/starting a dataflow or process group? Adding a template to the canvas? (if you are still using templates, they were deprecated in NiFi 1.x and Flow definitions are the new method to use. Templates were officially removed in NiFi 2.x). I have corruption issues related to templates previously. When you restarted your service, what observations were made in the nifi-app.log on all three nodes during startup? A flow election happens first where like flows each get a vote, the flow with the most ote becomes the cluster flow and nodes without that flow will join and inherit that cluster flow. One of your three nodes would have been elected as cluster coordinator and all other nodes would have formed the cluster by sending heartbeats to that node. During that node connection phase, any node with a mismatched flow would inherit the cluster flow. Any logging related to one or more of your nodes inheriting the cluster flow on startup? If not that then could be possible that some component was stuck in an enabling component state. So when you start a component(s) on the canvas, the component goes initially to "starting" and then "started". Likewise, stopping a component transitions to "stopping" and then "stopped". You may have been in situation where your nodes had a component stuck in the "stopping" or "starting" phase, but your cluster coordinator completed the transition. This could be caused by a bug in the component, load on the system, component has very long running process or hung process working on a FlowFile with large content, etc... Inspecting some thread dumps from those disconnected running nodes might help identifying scenario. This might be the most likely cause for you? I say this because if your flow.json.gz was corrupt, restarting your cluster would have had exceptions when trying to load the corrupted flow.json.gz. When stopping the nodes, NiFi eventually times out waiting for nodes to gracefully completed running threads and kills them. Then on restart no flow.json.gz corruption, all nodes restart fine, flow loaded successfully, and set the components to same running state. While none of above is a definitive answer because there is not enough info to provide that, hopefully this give you an idea of what could have happened so you know what to collect or look at deeper should it happen again. I will add that a number of fixes have gone into the newer releases (some around NiFi clustering). Apache NIFi 1.x is officially end of life. If you can not migrate to the newer Apache NiFi 2.x branch, you should at least upgrade to the latest Apache NiFi 1.28 release to take advantage of fixes done there since 1.23. Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
03-11-2026
05:39 AM
1 Kudo
@oka Can you share more details about your Apache NiFi version and dataflow configuration? Full Apache NiFi version PublishJMS processor configuration Sample FlowFile metadata/attributes of a FlowFile queued to the PublishJMS processor (exclude any sensitive values). Which FlowFile contains the attribute string you expect to be sent to JMS endpoint These details can be helpful in evaluating your setup and possible issue/solution. Thank you, Matt
... View more
03-05-2026
07:29 AM
@Frank168 I'd expect what you are seeing with the long sync times. Sync interval has not impact in speed of the user an group syncs. This simply controls how often NiFi is expected re-sync with ldap. So in your case, NiFi would finding one sync and immediately start the next. NiFi does a sync during startup as it loads the authorizers.xml, NiFi does not continue to load until that initial sync completes otherwise it can't do authorizations. After that initial successful sync with ldap, NiFi will finish loading and only then will UI become available. While running the sync interval will run in the background to re-sync from ldap in case any users or groups are added/removed/updated since last sync. some users adjust this to sync every 24 hours instead of every 30 minutes. This depends on how quickly you want NiFi to be aware of changes. As I commented before, you are loading 50,000+ users into NiFi heap memory (this impact the free heap available to the NiFi process for executing your dataflows on the canvas) and then redoing that every sync interval. This was not how NiFi ldap-user-group-provider was intended to be used. The only users and groups that should be loaded in to NiFi are those you plan to establish authorization policies for and will be accessing your secured NiFi. Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
03-04-2026
06:54 AM
@Krish98 Unfortunately, I have no "Best Practice" recommendation for the use case you have shared. It is not a use case I have ever setup before. Also want to share that in Apache NiFi 2.4+ version a new JWTBearerOAuth2AccessTokenProvider controller service was introduced. While not a solution to you query, I wanted to share this with you. Apache NiFi jira: NIFI-14380 NOTE: The Apache NiFi 1.x major release line is End-Of-Life now. There will be no future releases of the 1. major release line. There is no direct upgrade path from Apache NiFi 1.x to Apache NiFi 2.x. You'll need to migrate your dataflows from 1.x to 2.x. For our Cloudera Flow Management licensed users, we provide tooling to assist with migrating dataflows from Flow Management versions based on Apache NiFi 1.x to Flow Management versions based on Apache NiFi 2.x. Cloudera Flow Management 2.x also includes many of the components deprecated and no longer included in the Apache NiFi 2.x release line. Thank you, Matt
... View more
03-04-2026
06:36 AM
@Frank168 When NiFi fails to start, you'll see logging in the nifi-app.log with details as to why. My guess would be it failed when loading the authorizer because the ldap-user-group-provider execution was unsuccessful. That in turn was likely caused because you set the client page size to a value higher then the max enforced by the ldap server. LDAP server common MaxPageSize settings are 500 or 1000. Setting Page Size in NiFi to 500 is a common practice. The NiFi ldap client will request all returns in pages of 500. The Ldap server will return as many pages as required to fulfill the request. When page size is left blank in NiFi, the ldap server is only going to return one page limiting the result set to whatever fit in the ldap server page page size which would explain your missing returns. Asa reminder, I caution against returning so many users. Do you expect all 1700+ users to be accessing your NiFi? That would be extremely uncommon. Typical usage is that ldap users that will be accessing NiFi are added to a few specific ldap groups. Then the ldap-user-group-provider is configured to fetch only those groups and group members instead of syncing everyone from ldap. This strategy limits the heap consumed by the return user and group set. Keep in mind that the ldap-user-group-provider default is to re-sync every 30 minutes also. Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
03-03-2026
05:35 AM
@Frank168 What you expect is exactly how it works. Did you configure any user search based filtering in your LdapuserGroupProvider? Is your Referral Strategy set to FOLLOW? Any exceptions in the nifi-user.log or nifi-app.log? How many users are being returned successfully (count)? Did you set your Page Size? Ldap often limits the number of returns in a single page. NiFi by default does not set a "Page Size" so it request all results in one page. So depending on how many users you are trying to return and your ldap's defaults, this could result in missing users. Try setting the "Page Size" in NiFi to 500 to see if that resolves your issue. NOTE: All user and group identities returned by the ldap-user-group-provider are held in NiFi's heap memory. You should avoid syncing very large sets of users and groups and utilize user and group search filtering to only return those users and groups that will be accessing NiFi. If page size is not your issue, inspecting your setup may provide other clues. Can you share your login-identity-providers.xml and authorizers.xml file configurations (redact sensitive values)? Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
02-18-2026
09:59 AM
@rajivswe_2k7 Any user that has "modify the component" access granted to them on a process group will be able to upload a template via that process group. Users with that same authorization will also be able to use "add template" to add the uploaded template to the NiFi canvas within the authorized process group. There is no method built into Apache NiFi 1.x to disable templates or block user authorized as described above from being able to upload or add templates to canvas. It may make sense in some situations to upload a template to NiFi and then add it to the canvas so that it can be downloaded in the new flow definition format (Some template will require modification before they can be downloaded as a flow definition). Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
02-17-2026
06:14 AM
@hckorkmaz01 @vafs provided a couple excellent options already. As a third option, you could utilize the SiteToSiteStatusReportingTask. You can configure this NiFi reporting task to send metrics for only connections and you can add a regex to limited the number of connections returned (this requires renaming the desired connections so you have more granular control with the regex). This Reporting task can be configured to send to a remote input port on on the same NiFi. This Reporting task will execute the query on each host in your NiFi Cluster and send results to the the desired Site-To-Site remote input port. So in a three node NiFi cluster three FlowFiles will be produced and sent with Json content for connection(s) you filtered on. This way you can see what is queued per node. Limitations to monitoring connections in this way exist. Keep in mind that you are grabbing a snapshot at one specific moment in time. So if the downstream component of this monitored connection is running, the connection stat would have likely changed by the time you got the last status details. Now if you are using this to monitor flow dead-ends where you don't expect FlowFiles to ever accumulate, that is perhaps a valid use case. Example json output placed in FlowFile: [ {
"statusId" : "27b01368-8280-4e5e-ac22-0e749c46fe17",
"timestampMillis" : 1771336855410,
"timestamp" : "2026-02-17T14:00:55.410Z",
"actorHostname" : "nifi-node-1",
"componentType" : "Connection",
"componentName" : "Monitored-success",
"parentId" : "603017c1-0197-1000-0000-000013171e7c",
"parentName" : "test",
"parentPath" : "NiFi Flow / test",
"platform" : "nifi",
"application" : "NiFi Flow",
"componentId" : "508a2293-019a-1000-ffff-fffff15dc582",
"sourceId" : "5089d4ab-019a-1000-ffff-ffff9b682ab3",
"sourceName" : "UpdateAttribute",
"destinationId" : "4dc23b89-25b0-1eff-b974-6fe5425c283f",
"destinationName" : "UpdateAttribute",
"maxQueuedBytes" : 0,
"maxQueuedCount" : 0,
"queuedBytes" : 30720,
"queuedCount" : 30,
"inputBytes" : 0,
"inputCount" : 0,
"outputBytes" : 0,
"outputCount" : 0,
"backPressureBytesThreshold" : 1073741824,
"backPressureObjectThreshold" : 10000,
"backPressureDataSizeThreshold" : "1 GB",
"isBackPressureEnabled" : "false"
} ] IMPORTANT NOTE: I'd also like to point out that Apache NiFi 1.18 was released back in 2021 and newer releases have addressed many bugs and CVEs. I strongly encourage you to transition to a newer version of Apache NiFi. If you aren't ready to migrate to the newer Apache NiFi 2.x major release versions, you can still easily upgrade to last Apache NiFi 1.x major release version (1.28.0). Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
02-04-2026
10:03 AM
@zzzz77 Provenance can be very noisy depending on size of your dataflows and the amount of FlowFIles being processed through those dataflows. The provenance repo has age and size configuration that trigger roll-off of old events. So you may not reach the retention age if you reach size first. Also would not be trying to read provenance files while they are being written to. The SiteToSiteProvenanceReportingTask might be the solution you are looking for in Apache NiFi. This reporting task will send all provenance events over Site-To-Site protocol to a target NiFi where you can then feed them into any long term storage medium of your choice in a human readable format. Please help our community grow. If you found any of the suggestions/solutions provided helped you with solving your issue or answering your question, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more