<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Nifi node doesn't join the cluster anymore in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/237369#M199182</link>
    <description>&lt;P&gt;Ok so after increasing all the timeout parameter and thread for handling cluster protocol listed here :&lt;/P&gt;&lt;PRE&gt;nifi.zookeeper.connect.timeout
nifi.zookeeper.session.timeout
nifi.cluster.node.protocol.max.threads
nifi.cluster.node.protocol.threads
nifi.cluster.node.connection.timeout
nifi.cluster.node.read.timeout
nifi.cluster.protocol.heartbeat.interval&lt;/PRE&gt;&lt;P&gt;My node now start and can join the cluster but after a few minutes, I'm having the following bulletin&lt;/P&gt;&lt;PRE&gt;response time from was slow for each of the last 3 requests made&lt;/PRE&gt;&lt;P&gt;And the node start to behave badly and I need to stop it to stabilize the cluster.&lt;/P&gt;&lt;P&gt;Also it seems that this node have flowfile metadata which does not match the content repository and I have many log like the following&lt;/P&gt;&lt;PRE&gt;org.apache.nifi.processor.exception.FlowFileHandlingException: StandardFlowFileRecord[uuid=68f7abcf-fb70-4c8c-b0d1-4b6aaf64dc90,claim=,offset=0,name=91993057860031,size=0] is not known in this session (StandardProcessSession[id=336081
])&lt;/PRE&gt;&lt;P&gt;Does someone know a solution to purge all the flowfile while the node is offline ?&lt;/P&gt;</description>
    <pubDate>Fri, 09 Aug 2019 19:54:19 GMT</pubDate>
    <dc:creator>p_vigreux</dc:creator>
    <dc:date>2019-08-09T19:54:19Z</dc:date>
    <item>
      <title>Nifi node doesn't join the cluster anymore</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/237368#M199181</link>
      <description>&lt;P&gt;Hello Everyone,&lt;/P&gt;&lt;P&gt;We have a Nifi production cluster made of 4 nodes from HDF 3.1.2.0, using our HDP Zookeeper. We had a major breakdown and had to restart all cluster and since the restart one of the node is unable to join the cluster again. &lt;/P&gt;&lt;P&gt;Here is the log I have on it :&lt;/P&gt;&lt;P&gt;2019-08-06 17:56:32,385 ERROR [Curator-Framework-0] o.a.c.f.imps.CuratorFrameworkImpl Background retry gave up&lt;BR /&gt;org.apache.curator.CuratorConnectionLossException: KeeperErrorCode = ConnectionLoss&lt;BR /&gt;        at org.apache.curator.framework.imps.CuratorFrameworkImpl.performBackgroundOperation(CuratorFrameworkImpl.java:838)&lt;BR /&gt;        at org.apache.curator.framework.imps.CuratorFrameworkImpl.backgroundOperationsLoop(CuratorFrameworkImpl.java:809)&lt;BR /&gt;        at org.apache.curator.framework.imps.CuratorFrameworkImpl.access$300(CuratorFrameworkImpl.java:64)&lt;BR /&gt;        at org.apache.curator.framework.imps.CuratorFrameworkImpl$4.call(CuratorFrameworkImpl.java:267)&lt;BR /&gt;        at java.util.concurrent.FutureTask.run(FutureTask.java:266)&lt;BR /&gt;        at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)&lt;BR /&gt;        at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)&lt;BR /&gt;        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)&lt;BR /&gt;        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)&lt;BR /&gt;        at java.lang.Thread.run(Thread.java:748)&lt;BR /&gt;2019-08-06 17:56:32,394 INFO [main-EventThread] o.a.c.f.state.ConnectionStateManager State change: LOST&lt;BR /&gt;2019-08-06 17:56:32,394 INFO [Curator-ConnectionStateManager-0] o.a.n.c.l.e.CuratorLeaderElectionManager org.apache.nifi.controller.leader.election.CuratorLeaderElectionManager$ElectionListener@1ef0f788 Connection State changed to LOST&lt;BR /&gt;2019-08-06 17:56:32,394 INFO [Curator-ConnectionStateManager-0] o.a.n.c.l.e.CuratorLeaderElectionManager org.apache.nifi.controller.leader.election.CuratorLeaderElectionManager$ElectionListener@10371a26 Connection State changed to LOST&lt;BR /&gt;2019-08-06 17:56:32,398 INFO [main-EventThread] o.a.c.f.state.ConnectionStateManager State change: RECONNECTED&lt;BR /&gt;2019-08-06 17:56:32,398 INFO [Curator-ConnectionStateManager-0] o.a.n.c.l.e.CuratorLeaderElectionManager org.apache.nifi.controller.leader.election.CuratorLeaderElectionManager$ElectionListener@1ef0f788 Connection State changed to RECONNECTED&lt;BR /&gt;2019-08-06 17:56:32,398 INFO [Curator-ConnectionStateManager-0] o.a.n.c.l.e.CuratorLeaderElectionManager org.apache.nifi.controller.leader.election.CuratorLeaderElectionManager$ElectionListener@10371a26 Connection State changed to RECONNECTED&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;When trying a manual connection with the zkCli I don't have any issue and can browse the znode. I try to restart all the cluster several time, and even reboot the server. But for now no success except for a couple of hours and then the node is disconnected again and giving the same log.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Any idea would be wonderful ?&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Best regards&lt;/P&gt;</description>
      <pubDate>Wed, 07 Aug 2019 20:42:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/237368#M199181</guid>
      <dc:creator>p_vigreux</dc:creator>
      <dc:date>2019-08-07T20:42:01Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi node doesn't join the cluster anymore</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/237369#M199182</link>
      <description>&lt;P&gt;Ok so after increasing all the timeout parameter and thread for handling cluster protocol listed here :&lt;/P&gt;&lt;PRE&gt;nifi.zookeeper.connect.timeout
nifi.zookeeper.session.timeout
nifi.cluster.node.protocol.max.threads
nifi.cluster.node.protocol.threads
nifi.cluster.node.connection.timeout
nifi.cluster.node.read.timeout
nifi.cluster.protocol.heartbeat.interval&lt;/PRE&gt;&lt;P&gt;My node now start and can join the cluster but after a few minutes, I'm having the following bulletin&lt;/P&gt;&lt;PRE&gt;response time from was slow for each of the last 3 requests made&lt;/PRE&gt;&lt;P&gt;And the node start to behave badly and I need to stop it to stabilize the cluster.&lt;/P&gt;&lt;P&gt;Also it seems that this node have flowfile metadata which does not match the content repository and I have many log like the following&lt;/P&gt;&lt;PRE&gt;org.apache.nifi.processor.exception.FlowFileHandlingException: StandardFlowFileRecord[uuid=68f7abcf-fb70-4c8c-b0d1-4b6aaf64dc90,claim=,offset=0,name=91993057860031,size=0] is not known in this session (StandardProcessSession[id=336081
])&lt;/PRE&gt;&lt;P&gt;Does someone know a solution to purge all the flowfile while the node is offline ?&lt;/P&gt;</description>
      <pubDate>Fri, 09 Aug 2019 19:54:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/237369#M199182</guid>
      <dc:creator>p_vigreux</dc:creator>
      <dc:date>2019-08-09T19:54:19Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi node doesn't join the cluster anymore</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/270283#M207397</link>
      <description>&lt;P&gt;We end this by completely re-installing the failing node.&lt;/P&gt;</description>
      <pubDate>Thu, 12 Sep 2019 07:34:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-node-doesn-t-join-the-cluster-anymore/m-p/270283#M207397</guid>
      <dc:creator>p_vigreux</dc:creator>
      <dc:date>2019-09-12T07:34:20Z</dc:date>
    </item>
  </channel>
</rss>

