<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance. in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334269#M231691</link>
    <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/69192"&gt;@DigitalPlumber&lt;/a&gt;&amp;nbsp;The state file in local were deleted and the cluster is showing all the 3 nodes, thanks for the help.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;However, we are trying fault tolerance feature of NiFi, and when primary node goes down can see this message in other nodes:&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Cannot replicate request to Node nifi-hatest-03:9091 because the node is not connected&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;Below is article which had same errors, but we cannot clean state file always as if one node goes down our NiFi cluster should be still available without manual intervention?&lt;BR /&gt;&lt;A href="https://community.cloudera.com/t5/Support-Questions/NiFi-Cannot-replicate-request-to-Node-nifi-domain-net-8443/td-p/283664" target="_blank"&gt;https://community.cloudera.com/t5/Support-Questions/NiFi-Cannot-replicate-request-to-Node-nifi-domain-net-8443/td-p/283664&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 20 Jan 2022 05:39:40 GMT</pubDate>
    <dc:creator>chitrarthasur</dc:creator>
    <dc:date>2022-01-20T05:39:40Z</dc:date>
    <item>
      <title>NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334221#M231657</link>
      <description>&lt;P&gt;Hello Guys,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have 3 node NiFi cluster and the NiFi version is 1.12.1&lt;/P&gt;&lt;P&gt;nifi-hatest-01&lt;BR /&gt;nifi-hatest-02&lt;BR /&gt;nifi-hatest-03&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;All the 3 nodes are in cluster, but not in UI.&lt;/P&gt;&lt;P&gt;In all the 3 NiFi nodes, we see this 2 nodes in cluster mode:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="chitrarthasur_0-1642608010376.png" style="width: 848px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/33205iEDC0706274A61CD1/image-dimensions/848x126?v=v2" width="848" height="126" role="button" title="chitrarthasur_0-1642608010376.png" alt="chitrarthasur_0-1642608010376.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Now, as we see&amp;nbsp;nifi-hatest-02 is the primary and cooridantor. And&amp;nbsp;nifi-hatest-03 is not being shown here.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In&amp;nbsp;nifi-hatest-01, we see&amp;nbsp;nifi-hatest-01 is connecting to&amp;nbsp;nifi-hatest-02&lt;/P&gt;&lt;P&gt;2022-01-19 16:02:09,840 INFO [Clustering Tasks Thread-1] o.a.n.c.c.ClusterProtocolHeartbeater Heartbeat created at 2022-01-19 16:02:09,820 and sent to nifi-hatest-02:9088 at 2022-01-19 16:02:09,840; send took 20 millis&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In&amp;nbsp;nifi-hatest-03, we see&amp;nbsp;nifi-hatest-03 is also connecting to&amp;nbsp;nifi-hatest-02&lt;/P&gt;&lt;P&gt;2022-01-19 16:02:09,840 INFO [Clustering Tasks Thread-1] o.a.n.c.c.ClusterProtocolHeartbeater Heartbeat created at 2022-01-19 16:02:09,820 and sent to nifi-hatest-02:9088 at 2022-01-19 16:02:09,840; send took 20 millis&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In&amp;nbsp;nifi-hatest-02, we see&amp;nbsp;nifi-hatest-02 is also connecting to&amp;nbsp;all the 3 nodes as its coordinator:&lt;/P&gt;&lt;P&gt;2022-01-19 16:04:52,573 INFO [Process Cluster Protocol Request-30] o.a.n.c.p.impl.SocketProtocolListener Finished processing request 35f2ed1a-ca6f-4cc6-ab4a-6c0774fc9c6d (type=HEARTBEAT, length=2837 bytes) from nifi-hatest-03:9091 in 19 millis&lt;BR /&gt;2022-01-19 16:04:52,782 INFO [Process Cluster Protocol Request-31] o.a.n.c.p.impl.SocketProtocolListener Finished processing request 73d477cd-a936-4541-a854-03712cbc5fe9 (type=HEARTBEAT, length=2836 bytes) from nifi-hatest-02:9091 in 18 millis&lt;BR /&gt;2022-01-19 16:04:52,782 INFO [Clustering Tasks Thread-3] o.a.n.c.c.ClusterProtocolHeartbeater Heartbeat created at 2022-01-19 16:04:52,762 and sent to nifi-hatest-02:9088 at 2022-01-19 16:04:52,782; send took 19 millis&lt;BR /&gt;2022-01-19 16:04:53,916 INFO [Process Cluster Protocol Request-32] o.a.n.c.p.impl.SocketProtocolListener Finished processing request 4fbc39d9-5154-4841-b3be-ddde58b22cd2 (type=HEARTBEAT, length=2836 bytes) from nifi-hatest-01:9091 in 19 millis&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;No errors regarding cluster in either of nodes. But still see 2 nodes only in UI.&lt;/P&gt;&lt;P&gt;Also Cluster is using zk as cluster state provider.&lt;/P&gt;&lt;P&gt;Another thing is that when we create a processor in any of the NiFi UI, it gets propagated to other nodes too! So functionality its kind of working.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Jan 2022 16:22:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334221#M231657</guid>
      <dc:creator>chitrarthasur</dc:creator>
      <dc:date>2022-01-19T16:22:49Z</dc:date>
    </item>
    <item>
      <title>Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334230#M231662</link>
      <description>&lt;P&gt;This to me sounds like a hostname issue.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;could you confirm the value on nifi.properites for:&lt;/P&gt;&lt;P&gt;nifi.cluster.node.address=&lt;/P&gt;&lt;P&gt;nifi.web.https.host=&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Those values should match for the name of the host.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If nothing there stands out check log entries for latest messages contain below strings:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"org.apache.nifi.cluster.coordination.node.NodeClusterCoordinator: Status"&lt;/P&gt;&lt;P&gt;"org.apache.nifi.controller.StandardFlowService: Setting Flow Controller's Node ID:"&lt;/P&gt;&lt;P&gt;"org.apache.nifi.web.server.HostHeaderHandler"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Jan 2022 17:35:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334230#M231662</guid>
      <dc:creator>DigitalPlumber</dc:creator>
      <dc:date>2022-01-19T17:35:56Z</dc:date>
    </item>
    <item>
      <title>Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334231#M231663</link>
      <description>&lt;P&gt;Reading your sample log messages closer I can see that the coordinator received a heartbeat from node 3&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"&lt;SPAN&gt;2022-01-19 16:04:52,573 INFO [Process Cluster Protocol Request-30] o.a.n.c.p.impl.SocketProtocolListener Finished processing request 35f2ed1a-ca6f-4cc6-ab4a-6c0774fc9c6d (type=HEARTBEAT, length=2837 bytes) from nifi-hatest-03:9091 in 19 millis"&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;So I also wonder if we have a caching rendering&amp;nbsp;issue, can you see what the UI shows using incognito mode?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;And finally if this is a DEV environment you can also try and delete your local state directory, that value is set in file "state-management.xml"&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Deleting state will clear out any local state your processors might depend on if configured as such, so remove with caution.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;It will also&amp;nbsp;clear out cluster node ID's it local knows of.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Jan 2022 18:03:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334231#M231663</guid>
      <dc:creator>DigitalPlumber</dc:creator>
      <dc:date>2022-01-19T18:03:29Z</dc:date>
    </item>
    <item>
      <title>Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334263#M231688</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/69192"&gt;@DigitalPlumber&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have removed all these files from all the 3 nodes, as we don't have any flows.&lt;/P&gt;&lt;P&gt;sudo rm -f&amp;nbsp;/hadoop/nifi/conf/flow.xml.gz&lt;/P&gt;&lt;P&gt;sudo rm -f /hadoop/nifi/conf/users.xml&lt;BR /&gt;sudo rm -f /hadoop/nifi/conf/authorizations.xml&lt;BR /&gt;sudo rm -f /etc/nifi/3.5.2.0-99/0/state-management.xml&lt;BR /&gt;sudo rm -f /etc/nifi/3.5.2.0-99/0/authorizers.xml&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;And since its external zk, we are giving new root node to start with. But the issue still persists.&lt;/P&gt;&lt;P&gt;The new root node is "nifi3" and here is what i see in zk, is it expected output?&lt;/P&gt;&lt;P&gt;[zk: nifi-hatest-03:2181(CONNECTED) 2] ls /nifi3/leaders&lt;BR /&gt;[Primary Node, Cluster Coordinator]&lt;BR /&gt;[zk: nifi-hatest-03:2181(CONNECTED) 3] ls /nifi3/leaders/Primary Node&lt;BR /&gt;Node does not exist: /nifi3/leaders/Primary&lt;BR /&gt;[zk: nifi-hatest-03:2181(CONNECTED) 4] ls /nifi3/leaders/Cluster Coordinator&lt;BR /&gt;Node does not exist: /nifi3/leaders/Cluster&lt;BR /&gt;[zk: nifi-hatest-03:2181(CONNECTED) 5] ls /nifi3/components&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have also verified these 2 props and it matches the respective hostnames:&lt;/P&gt;&lt;P&gt;nifi.cluster.node.address=&lt;/P&gt;&lt;P&gt;nifi.web.https.host=&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 20 Jan 2022 03:38:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334263#M231688</guid>
      <dc:creator>chitrarthasur</dc:creator>
      <dc:date>2022-01-20T03:38:31Z</dc:date>
    </item>
    <item>
      <title>Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334269#M231691</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/69192"&gt;@DigitalPlumber&lt;/a&gt;&amp;nbsp;The state file in local were deleted and the cluster is showing all the 3 nodes, thanks for the help.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;However, we are trying fault tolerance feature of NiFi, and when primary node goes down can see this message in other nodes:&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Cannot replicate request to Node nifi-hatest-03:9091 because the node is not connected&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;Below is article which had same errors, but we cannot clean state file always as if one node goes down our NiFi cluster should be still available without manual intervention?&lt;BR /&gt;&lt;A href="https://community.cloudera.com/t5/Support-Questions/NiFi-Cannot-replicate-request-to-Node-nifi-domain-net-8443/td-p/283664" target="_blank"&gt;https://community.cloudera.com/t5/Support-Questions/NiFi-Cannot-replicate-request-to-Node-nifi-domain-net-8443/td-p/283664&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 20 Jan 2022 05:39:40 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334269#M231691</guid>
      <dc:creator>chitrarthasur</dc:creator>
      <dc:date>2022-01-20T05:39:40Z</dc:date>
    </item>
    <item>
      <title>Re: NiFi Node showing 2 nodes and not respecting node down fault tolerance.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334308#M231709</link>
      <description>&lt;P&gt;Deleting that state directory should not be a normal maintenance function.&lt;/P&gt;&lt;P&gt;What you initially described is a very odd case.&lt;/P&gt;&lt;P&gt;The fact that your node went down 003 and if it was the primary node or the coordinator node, internally there would have been a election to nominate a new cluster member node to perform those functions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In your case node 003 is a member of the cluster but it is not connected.&lt;/P&gt;&lt;P&gt;Why it is not connected could be the cause of n reasons typically node is down or it was manually disconnected. When you see that message how many member nodes do you have?&lt;BR /&gt;I expect the UI to show 2/3 because node 3 is not connected.&lt;BR /&gt;The solution is to connect it by fixing the issue of why node is down or connect it through the UI&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 20 Jan 2022 17:43:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/NiFi-Node-showing-2-nodes-and-not-respecting-node-down-fault/m-p/334308#M231709</guid>
      <dc:creator>DigitalPlumber</dc:creator>
      <dc:date>2022-01-20T17:43:22Z</dc:date>
    </item>
  </channel>
</rss>

