<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question CDH Express 6.3.1 - Can't start HDFS after cluster crash and apparently not clean namenode failover in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/CDH-Express-6-3-1-Can-t-start-HDFS-after-cluster-crash-and/m-p/395525#M248964</link>
    <description>&lt;P&gt;Hi, here is a summary of our blocking situation on our CDH Express 6.3.1 installation:&lt;/P&gt;&lt;P&gt;- A crash happened on the storage where all the virtual nodes composing our Cluster insist&lt;/P&gt;&lt;P&gt;- After solving the issue on the storage, our Cloudera cluster started again working, but with a critical error on HDFS Service Nodename&lt;/P&gt;&lt;P&gt;- In fact, we have High Availability configured on 2 nodes, and exploring the situation we saw that Namenode 2 was active and Namenode 1 was DOWN&lt;/P&gt;&lt;P&gt;- But the cluster worked&lt;/P&gt;&lt;P&gt;- In an attempt to fix the issue, we tried several times to manually start the Namenode 1 from "HDFS --&amp;gt; Instances --&amp;gt; federation --&amp;gt; Namenode1 --&amp;gt; Start" but it was impossible to start manually&lt;/P&gt;&lt;P&gt;- So, we decided to disable High Availability but the new configuration has not completed and actually is in pending status&lt;/P&gt;&lt;P&gt;- It's in fact impossible to start HDFS service at all because an error pops up saying that "&lt;SPAN&gt;&lt;STRONG&gt;Nameservice nameservice1 has no SecondaryNameNode or High-Availability partner&lt;/STRONG&gt;"&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;It is interesting to note that (see attached picture) apparently we don't have High Availability configured anymore but cluster does not start anyhow&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;As you can see from the second screenshot, if we try to manually start HDFS service a configuration issue is shown and it still believes to be running in High Availability mode and still looking for a "partner node"&lt;/P&gt;&lt;P&gt;QUESTION:&lt;/P&gt;&lt;P&gt;- Is there a way to force a NON-High-availability state and start the HDFS Service?&lt;/P&gt;&lt;P&gt;Thank you for any advice and/or insights&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="immagine (1).png" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/42237iB30537E087ABB860/image-size/large?v=v2&amp;amp;px=999" role="button" title="immagine (1).png" alt="immagine (1).png" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="immagine.png" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/42238iD3BB45CE5D9A9E71/image-size/large?v=v2&amp;amp;px=999" role="button" title="immagine.png" alt="immagine.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
    <pubDate>Fri, 18 Oct 2024 16:58:34 GMT</pubDate>
    <dc:creator>FrozenWave</dc:creator>
    <dc:date>2024-10-18T16:58:34Z</dc:date>
    <item>
      <title>CDH Express 6.3.1 - Can't start HDFS after cluster crash and apparently not clean namenode failover</title>
      <link>https://community.cloudera.com/t5/Support-Questions/CDH-Express-6-3-1-Can-t-start-HDFS-after-cluster-crash-and/m-p/395525#M248964</link>
      <description>&lt;P&gt;Hi, here is a summary of our blocking situation on our CDH Express 6.3.1 installation:&lt;/P&gt;&lt;P&gt;- A crash happened on the storage where all the virtual nodes composing our Cluster insist&lt;/P&gt;&lt;P&gt;- After solving the issue on the storage, our Cloudera cluster started again working, but with a critical error on HDFS Service Nodename&lt;/P&gt;&lt;P&gt;- In fact, we have High Availability configured on 2 nodes, and exploring the situation we saw that Namenode 2 was active and Namenode 1 was DOWN&lt;/P&gt;&lt;P&gt;- But the cluster worked&lt;/P&gt;&lt;P&gt;- In an attempt to fix the issue, we tried several times to manually start the Namenode 1 from "HDFS --&amp;gt; Instances --&amp;gt; federation --&amp;gt; Namenode1 --&amp;gt; Start" but it was impossible to start manually&lt;/P&gt;&lt;P&gt;- So, we decided to disable High Availability but the new configuration has not completed and actually is in pending status&lt;/P&gt;&lt;P&gt;- It's in fact impossible to start HDFS service at all because an error pops up saying that "&lt;SPAN&gt;&lt;STRONG&gt;Nameservice nameservice1 has no SecondaryNameNode or High-Availability partner&lt;/STRONG&gt;"&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;It is interesting to note that (see attached picture) apparently we don't have High Availability configured anymore but cluster does not start anyhow&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;As you can see from the second screenshot, if we try to manually start HDFS service a configuration issue is shown and it still believes to be running in High Availability mode and still looking for a "partner node"&lt;/P&gt;&lt;P&gt;QUESTION:&lt;/P&gt;&lt;P&gt;- Is there a way to force a NON-High-availability state and start the HDFS Service?&lt;/P&gt;&lt;P&gt;Thank you for any advice and/or insights&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="immagine (1).png" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/42237iB30537E087ABB860/image-size/large?v=v2&amp;amp;px=999" role="button" title="immagine (1).png" alt="immagine (1).png" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="immagine.png" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/42238iD3BB45CE5D9A9E71/image-size/large?v=v2&amp;amp;px=999" role="button" title="immagine.png" alt="immagine.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 18 Oct 2024 16:58:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/CDH-Express-6-3-1-Can-t-start-HDFS-after-cluster-crash-and/m-p/395525#M248964</guid>
      <dc:creator>FrozenWave</dc:creator>
      <dc:date>2024-10-18T16:58:34Z</dc:date>
    </item>
    <item>
      <title>Re: CDH Express 6.3.1 - Can't start HDFS after cluster crash and apparently not clean namenode failover</title>
      <link>https://community.cloudera.com/t5/Support-Questions/CDH-Express-6-3-1-Can-t-start-HDFS-after-cluster-crash-and/m-p/395534#M248970</link>
      <description>&lt;P&gt;Solved by:&lt;/P&gt;&lt;P&gt;- Deleting the Namenode role entirely from "HDFS --&amp;gt; Instances --&amp;gt; NameNode", clicking the checkbox near the Namenode Instance and selecting "delete" from the "Actions" dropdown menu&lt;/P&gt;&lt;P&gt;- Redeploying a new "Namenode" role on the same host where the primary Namenode was previously running&lt;/P&gt;&lt;P&gt;- Reenabling High Availability on the Namenode&lt;/P&gt;&lt;P&gt;- Reconstructing the Namenode Metadata in the "Hive" service, under "Hive --&amp;gt; actions --&amp;gt; rebuild Namenode metadata"&lt;/P&gt;</description>
      <pubDate>Fri, 18 Oct 2024 21:15:05 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/CDH-Express-6-3-1-Can-t-start-HDFS-after-cluster-crash-and/m-p/395534#M248970</guid>
      <dc:creator>FrozenWave</dc:creator>
      <dc:date>2024-10-18T21:15:05Z</dc:date>
    </item>
  </channel>
</rss>

