Member since
07-17-2019
738
Posts
433
Kudos Received
111
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1971 | 08-06-2019 07:09 PM | |
2268 | 07-19-2019 01:57 PM | |
2996 | 02-25-2019 04:47 PM | |
3505 | 10-11-2018 02:47 PM | |
1032 | 09-26-2018 02:49 PM |
01-10-2017
07:45 PM
1 Kudo
When executing Step 3 of the Ambari installation wizard "Confirm Hosts", Ambari will (by default) SSH to each node and start an instance of the Ambari Agent process. In some cases, it is possible that the local RPM database is corrupted and this registration process will fail. The error message in Ambari would look something like: INFO:root:Executing parallel bootstrap
ERROR:root:ERROR: Bootstrap of host myhost.mydomain fails because previous action finished with non-zero exit code (1)
ERROR MESSAGE: tcgetattr: Invalid argumentConnection to myhost.mydomain closed.
STDOUT: Error: database disk image is malformed
Error: database disk image is malformedDesired version (2.5.0.0) of ambari-agent package is not available.
tcgetattr: Invalid argumentConnection to myhost.mydomain closed. In this case, the local RPM database is malformed and all actions to alter the installed packages on the system will fail until the database is rebuilt. This can be done by the following commands as root on the host reporting the error: [root@myhost ~] # mv /var/lib/rpm/__db* /tmp
[root@myhost ~] # rpm --rebuilddb Then, click the "Retry Failed Hosts" button in Ambari and the registration should succeed.
... View more
Labels:
01-05-2017
05:25 PM
It would appear that your DataNode is failing which is the cause of the other services failing. It also appears that you have not changed the default hdfs-site.xml configuration that controls where DataNodes store their data on the local filesystem. It is not uncommon for operation systems to wipe the /tmp directory (on boot). Perhaps you have experienced this and need to re-format your HDFS? Change dfs.datanode.data.dir, dfs.namenode.name.dir, and dfs.namenode.checkpoint.dir, then format HDFS $ hdfs namenode -format Beware: Formatting HDFS is a destructive operation. Do not perform this operation unless all of the data in HDFS is stored elsewhere or can be generated.
... View more
01-03-2017
08:52 PM
So I guess you didn't have hostname resolution set up correctly as you said below? 🙂 But in general, yes, all host advertisements done by HBase are done using hostnames and not IP addresses. This is essentially a prerequisite to get Kerberos authentication working.
... View more
01-03-2017
04:12 PM
Did you inspect the extra logging at the client side? It looks like you have only copied the HBase master server logs. Also, what ports did you verify via telnet?
... View more
01-03-2017
02:20 AM
Please re-read the description on HBASE-14729. There no code-changes made by that JIRA issue -- it was closed as a duplicate of https://issues.apache.org/jira/browse/HBASE-14223 which is still outstanding.
... View more
12-24-2016
05:37 PM
1 Kudo
I would guess the problem does not lie in your client and ZooKeeper, but your client and HBase. Remember that one use of ZooKeeper is for discovering HBase servers. I would verify that the service ports for HBase (e.g. 16000, 16020) are bound to an external network interface (*not* lo or 127.0.0.1) using netstat and that you can connect to these ports remotely using telnet as you did. 16000 is the RPC port for the Master and 16020 is the RPC for the RegionServer. Another option to get more debug information is to increase the log verbosity to DEBUG via log4j in your client for the org.apache.hadoop.hbase and org.apache.phoenix packages. This should give you more information about what actions the client is taking and why they are failing.
... View more
12-20-2016
06:14 PM
Hah, yes, it seems like you have a port conflict problem.
You could use a tool like netstat to find what process has already bound the port 60020, e.g. `sudo netstat -nape | fgrep 60020`. You can find the pid of the process which has that port bound. Once you identify the other process, you can determine if there is a port conflict which needs to be changed via configuration.
One important note is that 60020 is in the Ephemeral port range which means that there may be transient sockets binding that port. If you do not see any service bound on that port now, this is likely what happened. You can try to just restart the AMS in this case. This is the reason that HBase default ports moved from 600xx to 160xx in recent versions.
... View more
12-20-2016
04:50 PM
@ARUN Can you please share the error that you see? I assume this is from the AMS log files.
... View more
12-19-2016
02:50 PM
Yes, the default is "true". That's why I stated my reply in the way I did. As long as you are not setting the property to false, the (default) value would be true, and thus the table would not be disabled in the process.
... View more
12-17-2016
08:05 PM
No, you are incorrect, Sami. HBase knows the set of column families. It does *not* track the set of qualifiers.
... View more