Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Error add node into cluster

Solved Go to solution
Highlighted

Error add node into cluster

New Contributor

Folks, After spend a little days trying install my fisrt ambari-server, I decided ask a little help :)

I have done all requeried procedures into installation guide but, when add a node, de process stay in preparing.

It is normal spend a lot of time to do ? I stayed 24 hours waiting the the process finish but not happened.

The follow the the imagens of configuration procedure.

https://i.imgur.com/aAVCSzV.png

https://i.imgur.com/gPbVR4U.png

https://i.imgur.com/M99Jugd.png

https://i.imgur.com/UMcB3BG.png

The previus configuration (done):

SSH PublicKey - OK

Disable Firewall - OK

Disable Iptables - OK

Include ip into hosts file - OK

Install amabi-server - OK

Install amabi-agent - OK

Someone can give me some tip or trick about what i do wrong? Or how i can solve my problem.

Thanks so much folks!

The unique message that are showed into log:

tail server log:

31 May 2018 14:23:54,071  INFO [ambari-client-thread-34] AmbariMetaInfo:1449 - Stack HDP-2.3 is not active, skipping VDF
31 May 2018 14:23:54,071  INFO [ambari-client-thread-34] AmbariMetaInfo:1449 - Stack HDP-2.2 is not active, skipping VDF
31 May 2018 14:23:54,071  INFO [ambari-client-thread-34] AmbariMetaInfo:1449 - Stack HDP-2.1 is not active, skipping VDF
31 May 2018 14:23:54,071  INFO [ambari-client-thread-34] AmbariMetaInfo:1449 - Stack HDP-2.0 is not active, skipping VDF
31 May 2018 14:23:54,071  INFO [ambari-client-thread-34] AmbariMetaInfo:1449 - Stack BigInsights-4.2 is not active, skipping VDF
31 May 2018 14:24:01,847  INFO [Thread-23] AbstractPoolBackedDataSource:212 - Initializing c3p0 pool... com.mchange.v2.c3p0.ComboPooledDataSource [ acquireIncrement -> 3, acquireRetryAttempts -> 30, acquireRetryDelay -> 1000, autoCommitOnClose -> false, automaticTestTable -> null, breakAfterAcquireFailure -> false, checkoutTimeout -> 0, connectionCustomizerClassName -> null, connectionTesterClassName -> com.mchange.v2.c3p0.impl.DefaultConnectionTester, contextClassLoaderSource -> caller, dataSourceName -> z8kflt9v1ewa6kzjcpa2f|2a9f7572, debugUnreturnedConnectionStackTraces -> false, description -> null, driverClass -> org.postgresql.Driver, extensions -> {}, factoryClassLocation -> null, forceIgnoreUnresolvedTransactions -> false, forceSynchronousCheckins -> false, forceUseNamedDriverClass -> false, identityToken -> z8kflt9v1ewa6kzjcpa2f|2a9f7572, idleConnectionTestPeriod -> 50, initialPoolSize -> 3, jdbcUrl -> jdbc:postgresql://localhost/ambari, maxAdministrativeTaskTime -> 0, maxConnectionAge -> 0, maxIdleTime -> 0, maxIdleTimeExcessConnections -> 0, maxPoolSize -> 5, maxStatements -> 0, maxStatementsPerConnection -> 120, minPoolSize -> 1, numHelperThreads -> 3, preferredTestQuery -> select 0, privilegeSpawnedThreads -> false, properties -> {user=******, password=******}, propertyCycle -> 0, statementCacheNumDeferredCloseThreads -> 0, testConnectionOnCheckin -> true, testConnectionOnCheckout -> false, unreturnedConnectionTimeout -> 0, userOverrides -> {}, usesTraditionalReflectiveProxies -> false ]
31 May 2018 14:24:01,902  INFO [Thread-23] JobStoreTX:861 - Freed 0 triggers from 'acquired' / 'blocked' state.
31 May 2018 14:24:01,910  INFO [Thread-23] JobStoreTX:871 - Recovering 0 jobs that were in-progress at the time of the last shut-down.
31 May 2018 14:24:01,910  INFO [Thread-23] JobStoreTX:884 - Recovery complete.
31 May 2018 14:24:01,910  INFO [Thread-23] JobStoreTX:891 - Removed 0 'complete' triggers.
31 May 2018 14:24:01,911  INFO [Thread-23] JobStoreTX:896 - Removed 0 stale fired job entries.
31 May 2018 14:24:01,912  INFO [Thread-23] QuartzScheduler:575 - Scheduler ExecutionScheduler_$_NON_CLUSTERED started.
31 May 2018 14:26:35,107  INFO [ambari-client-thread-33] BootStrapImpl:108 - BootStrapping hosts localhost:
31 May 2018 14:26:35,122  INFO [Thread-31] BSRunner:189 - Kicking off the scheduler for polling on logs in /var/run/ambari-server/bootstrap/1
31 May 2018 14:26:35,122  INFO [Thread-31] BSRunner:372 - Error executing bootstrap Cannot create /var/run/ambari-server/bootstrap
31 May 2018 14:26:35,122 ERROR [Thread-31] BSRunner:441 - java.io.FileNotFoundException: /var/run/ambari-server/bootstrap/1/localhost.done (No such file or directory)
31 May 2018 14:26:35,123  WARN [Thread-31] BSRunner:401 - File does not exist: /var/run/ambari-server/bootstrap/1/sshKey
31 May 2018 14:27:01,846  INFO [pool-18-thread-1] MetricsServiceImpl:65 - Attempting to initialize metrics sink
31 May 2018 14:27:01,847  INFO [pool-18-thread-1] MetricsServiceImpl:81 - ********* Configuring Metric Sink **********
31 May 2018 14:27:01,847  INFO [pool-18-thread-1] AmbariMetricSinkImpl:95 - No clusters configured.
31 May 2018 14:32:01,848  INFO [pool-18-thread-1] MetricsServiceImpl:65 - Attempting to initialize metrics sink
31 May 2018 14:32:01,849  INFO [pool-18-thread-1] MetricsServiceImpl:81 - ********* Configuring Metric Sink **********
31 May 2018 14:32:01,849  INFO [pool-18-thread-1] AmbariMetricSinkImpl:95 - No clusters configured.
31 May 2018 14:37:01,853  INFO [pool-18-thread-1] MetricsServiceImpl:65 - Attempting to initialize metrics sink
31 May 2018 14:37:01,853  INFO [pool-18-thread-1] MetricsServiceImpl:81 - ********* Configuring Metric Sink **********
31 May 2018 14:37:01,853  INFO [pool-18-thread-1] AmbariMetricSinkImpl:95 - No clusters configured.
31 May 2018 14:42:01,855  INFO [pool-18-thread-1] MetricsServiceImpl:65 - Attempting to initialize metrics sink
31 May 2018 14:42:01,855  INFO [pool-18-thread-1] MetricsServiceImpl:81 - ********* Configuring Metric Sink **********
31 May 2018 14:42:01,856  INFO [pool-18-thread-1] AmbariMetricSinkImpl:95 - No clusters configured.
31 May 2018 14:47:01,859  INFO [pool-18-thread-1] MetricsServiceImpl:65 - Attempting to initialize metrics sink
31 May 2018 14:47:01,859  INFO [pool-18-thread-1] MetricsServiceImpl:81 - ********* Configuring Metric Sink **********
31 May 2018 14:47:01,859  INFO [pool-18-thread-1] AmbariMetricSinkImpl:95 - No clusters configured. 

tail agent log:

INFO 2018-05-31 14:22:58,839 NetUtil.py:70 - Connecting to https://edcserver.domain:8440/connection_info
INFO 2018-05-31 14:22:58,896 security.py:93 - SSL Connect being called.. connecting to the server
INFO 2018-05-31 14:22:58,938 security.py:60 - SSL connection established. Two-way SSL authentication is turned off on the server.
INFO 2018-05-31 14:22:59,011 Controller.py:196 - Registration Successful (response id = 0)
INFO 2018-05-31 14:22:59,012 AmbariConfig.py:316 - Updating config property (agent.check.remote.mounts) with value (false)
INFO 2018-05-31 14:22:59,012 AmbariConfig.py:316 - Updating config property (agent.auto.cache.update) with value (true)
INFO 2018-05-31 14:22:59,012 AmbariConfig.py:316 - Updating config property (java.home) with value (/usr/jdk64/jdk1.8.0_112)
INFO 2018-05-31 14:22:59,012 AmbariConfig.py:316 - Updating config property (agent.check.mounts.timeout) with value (0)
WARNING 2018-05-31 14:22:59,012 AlertSchedulerHandler.py:123 - There are no alert definition commands in the heartbeat; unable to update definitions
INFO 2018-05-31 14:22:59,012 Controller.py:516 - Registration response from edcserver.domain was OK
INFO 2018-05-31 14:22:59,012 Controller.py:521 - Resetting ActionQueue...
INFO 2018-05-31 14:23:09,016 Controller.py:304 - Heartbeat (response id = 0) with server is running...
INFO 2018-05-31 14:23:09,016 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:23:09,017 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:23:09,079 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:23:09,259 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:23:09,260 Controller.py:320 - Sending Heartbeat (id = 0)
INFO 2018-05-31 14:23:09,267 Controller.py:333 - Heartbeat response received (id = 1)
INFO 2018-05-31 14:23:09,268 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:23:09,268 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:23:09,268 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:23:09,268 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:23:09,268 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:23:19,169 Controller.py:482 - Wait for next heartbeat over
INFO 2018-05-31 14:24:18,661 Controller.py:304 - Heartbeat (response id = 7) with server is running...
INFO 2018-05-31 14:24:18,661 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:24:18,665 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:24:18,753 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:24:18,893 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:24:18,893 Controller.py:320 - Sending Heartbeat (id = 7)
INFO 2018-05-31 14:24:18,896 Controller.py:333 - Heartbeat response received (id = 8)
INFO 2018-05-31 14:24:18,896 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:24:18,896 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:24:18,896 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:24:18,896 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:24:18,896 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:24:28,797 Controller.py:482 - Wait for next heartbeat over
INFO 2018-05-31 14:25:28,275 Controller.py:304 - Heartbeat (response id = 14) with server is running...
INFO 2018-05-31 14:25:28,276 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:25:28,279 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:25:28,381 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:25:28,507 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:25:28,507 Controller.py:320 - Sending Heartbeat (id = 14)
INFO 2018-05-31 14:25:28,511 Controller.py:333 - Heartbeat response received (id = 15)
INFO 2018-05-31 14:25:28,511 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:25:28,511 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:25:28,511 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:25:28,511 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:25:28,511 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:25:38,412 Controller.py:482 - Wait for next heartbeat over
INFO 2018-05-31 14:26:37,884 Controller.py:304 - Heartbeat (response id = 21) with server is running...
INFO 2018-05-31 14:26:37,884 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:26:37,889 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:26:37,984 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:26:38,138 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:26:38,139 Controller.py:320 - Sending Heartbeat (id = 21)
INFO 2018-05-31 14:26:38,141 Controller.py:333 - Heartbeat response received (id = 22)
INFO 2018-05-31 14:26:38,141 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:26:38,142 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:26:38,142 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:26:38,142 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:26:38,142 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:26:48,043 Controller.py:482 - Wait for next heartbeat over
INFO 2018-05-31 14:27:47,512 Controller.py:304 - Heartbeat (response id = 28) with server is running...
INFO 2018-05-31 14:27:47,512 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:27:47,516 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:27:47,650 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:27:47,808 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:27:47,809 Controller.py:320 - Sending Heartbeat (id = 28)
INFO 2018-05-31 14:27:47,811 Controller.py:333 - Heartbeat response received (id = 29)
INFO 2018-05-31 14:27:47,811 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:27:47,811 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:27:47,811 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:27:47,811 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:27:47,811 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:27:57,714 Controller.py:482 - Wait for next heartbeat over
INFO 2018-05-31 14:28:57,189 Controller.py:304 - Heartbeat (response id = 35) with server is running...
INFO 2018-05-31 14:28:57,189 Controller.py:311 - Building heartbeat message
INFO 2018-05-31 14:28:57,193 Heartbeat.py:87 - Adding host info/state to heartbeat message.
INFO 2018-05-31 14:28:57,379 logger.py:75 - Testing the JVM's JCE policy to see it if supports an unlimited key length.
INFO 2018-05-31 14:28:57,558 Hardware.py:188 - Some mount points were ignored: /dev/shm, /run, /sys/fs/cgroup, /run/user/42, /run/user/0
INFO 2018-05-31 14:28:57,558 Controller.py:320 - Sending Heartbeat (id = 35)
INFO 2018-05-31 14:28:57,561 Controller.py:333 - Heartbeat response received (id = 36)
INFO 2018-05-31 14:28:57,561 Controller.py:342 - Heartbeat interval is 10 seconds
INFO 2018-05-31 14:28:57,561 Controller.py:380 - Updating configurations from heartbeat
INFO 2018-05-31 14:28:57,561 Controller.py:389 - Adding cancel/execution commands
INFO 2018-05-31 14:28:57,561 Controller.py:406 - Adding recovery commands
INFO 2018-05-31 14:28:57,561 Controller.py:475 - Waiting 9.9 for next heartbeat
INFO 2018-05-31 14:29:07,462 Controller.py:482 - Wait for next heartbeat over
1 ACCEPTED SOLUTION

Accepted Solutions
Highlighted

Re: Error add node into cluster

Mentor

@FELIPE CABRAL

Your /etc/hosts should look like this in the below example I used 192.168.0.124 as the IP of edcserver.domain please use the appropriate IP

The first 2 lines shouldn't be tampered with

 # cat /etc/hosts
127.0.0.1 localhost localhost.localdomain
::1 localhost localhost.localdomain
192.168.0.124  edcserver.domain  edcserver

Now to overcome the error 1 other registered host (screenshot 2018-05-31-18-09-36-window.png) you have to delete this entry from 2 tables hoststate and hosts in the Ambari database.

I have attached the procedure below, the assumption is as follows

  • Ambari database name = ambari
  • Ambari database type= MySQL
  • Ambari database password =Secret1234

If you are using postgres please adapt accordingly the SQL

Log on the DB

# mysql -u root -pSecret1234 

some output here

Check the databases

mysql> show databases; 
+--------------------+ 
| Database           | 
+--------------------+ 
| information_schema | 
| ambari             | 
| hive               | 
| mysql              | 
| oozie              | 
| ranger             | 
| rangerkms          | 
| test               | 
+--------------------+ 
8 rows in set (0.00 sec)

Connect to the Ambari database

mysql> use ambari; 
Reading table information for completion of table and column names 
You can turn off this feature to get a quicker startup with -A 
Database changed 

Check the structure if hoststate table we need to delete the host_id

mysql> describe hoststate; 
+-------------------+--------------+------+-----+---------+-------+ 
| Field | Type | Null | Key | Default | Extra | 
+-------------------+--------------+------+-----+---------+-------+ 
| agent_version | varchar(255) | NO | | NULL | | 
| available_mem | bigint(20) | NO | | NULL | | 
| current_state     | varchar(255) | NO   |     | NULL    |       | 
| health_status     | varchar(255) | YES  |     | NULL    |       | 
| host_id           | bigint(20)   | NO   | PRI | NULL    |       | 
| time_in_state     | bigint(20)   | NO   |     | NULL    |       | 
| maintenance_state | varchar(512) | YES  |     | NULL    |       | 
+-------------------+--------------+------+-----+---------+-------+ 
7 rows in set (0.03 sec) 
mysql> select host_id from hoststate; 
+---------+ 
| host_id | 
+---------+ 
| 1       | 
+---------+ 
1 row in set (0.04 sec) 
mysql> Delete from hoststate where host_id=1; 
Query OK, 1 rows affected (0.00 sec) 

Check the hosts table

mysql> describe hosts; 
+------------------------+---------------+------+-----+---------+-------+ 
| Field 		 | Type 	 | Null | Key | Default | Extra | 
+------------------------+---------------+------+-----+---------+-------+ 
| host_id 		 | bigint(20) 	 | NO 	| PRI | NULL 	| 	| 
| host_name 		 | varchar(255)  | NO 	| UNI | NULL 	| 	| 
| cpu_count 		 | int(11) 	 | NO 	|     | NULL    |       | 
| cpu_info  		 | varchar(255)  | NO   |     | NULL    |       | 
| discovery_status 	 | varchar(2000) | NO   |     | NULL    |       | 
| host_attributes 	 | longtext 	 | NO   |     | NULL    |       | 
| ipv4 			 | varchar(255)  | YES  |     | NULL    |       | 
| ipv6 			 | varchar(255)  | YES  |     | NULL    |       | 
| last_registration_time | bigint(20)    | NO   |     | NULL    |       | 
| os_arch 		 | varchar(255)  | NO   |     | NULL    |       | 
| os_info 		 | varchar(1000) | NO   |     | NULL    |       | 
| os_type 		 | varchar(255)  | NO   |     | NULL    |       | 
| ph_cpu_count 		 | int(11)       | YES  |     | NULL    |       | 
| public_host_name 	 | varchar(255)  | YES  |     | NULL    |       | 
| rack_info 		 | varchar(255)  | NO   |     | NULL    |       | 
| total_mem 		 | bigint(20)    | NO   |     | NULL    |       | 
+------------------------+---------------+------+-----+---------+-------+ 
16 rows in set (0.02 sec) 

Retrieve and delete the host with the matching id 1

mysql> select host_id from hosts; 
+---------+ 
| host_id | 
+---------+ 
| 1 	  | 
+---------+ 
1 row in set (0.00 sec) 
mysql> Delete from hoststate where host_id=1; 
Query OK, 1 rows affected (0.00 sec)

And restart ambari-server and proceed with the registration all should be fine.

Please revert If you found this answer addressed your question, please take a moment to log in and click the "accept" link on the answer.

View solution in original post

4 REPLIES 4
Highlighted

Re: Error add node into cluster

Mentor

@FELIPE CABRAL

Can you share the entry of your /etc/hosts on the Ambari server?

Select Version

Can you remove the repositories that doesn't match your node either the Redppc7 or Redhat

Install Option

Please use the FQDN rather than localhost, see the output of below command

# hostname -f

Please revert

Highlighted

Re: Error add node into cluster

New Contributor

@Geoffrey Shelton Okot thanks a lot by you attention

So... a stranger thing happen when I confere the result of the command that you ask to do. Differents names was showed to me.

I modified some things and now looks right.

Print screens

2018-05-31-18-03-10-window.png

2018-05-31-18-06-52-window.png

2018-05-31-18-09-07-window.png

2018-05-31-18-09-22-window.png

2018-05-31-18-09-36-window.png

How you can see, the modifications and new test failed. The same behavior continue happen.

You can see in the last print (2018-05-31-18-09-36-window.png) the locahost.localdomain was identified. I don't know where it can influence my process.

Existe some attention point that was done wrong? Just ambari-server and agent need be installed?

-----------------------------------------------------

Computer configuration

Centos 7

16gb ram

4 core

--------------------------------------------------------

Follow the command result and files that you ask:

─
[2018-05-31 17:57.23]  ~/Desktop/chaves
[felip.DESKTOP-183A5DG] ➤ ssh -i id_rsa_root root@192.168.15.205
Last login: Thu May 31 16:57:14 2018 from desktop-183a5dg
[root@edcserver ~]# hostname -f
edcserver
[root@edcserver ~]# hostnamectl status
   Static hostname: edcserver.domain
         Icon name: computer-vm
           Chassis: vm
        Machine ID: e53f426a62744fd7b2c4b7dada89308f
           Boot ID: db8f48d7be7a46249b7c65b224328300
    Virtualization: vmware
  Operating System: CentOS Linux 7 (Core)
       CPE OS Name: cpe:/o:centos:centos:7
            Kernel: Linux 3.10.0-862.3.2.el7.x86_64
      Architecture: x86-64
[root@edcserver ~]#

ambari-agent config:

#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific
[server]
hostname=edcserver.domain
url_port=8440
secured_url_port=8441
connect_retry_delay=10
max_reconnect_retry_delay=30
[agent]
logdir=/var/log/ambari-agent
piddir=/var/run/ambari-agent
prefix=/var/lib/ambari-agent/data
;loglevel=(DEBUG/INFO)
loglevel=INFO
data_cleanup_interval=86400

/etc/hosts:

[root@edcserver ~]# cat /etc/hosts
127.0.0.1 edcserver edcserver.domain localhost localhost.localdomain
::1 edcserver edcserver.domain localhost localhost.localdomain

Highlighted

Re: Error add node into cluster

Mentor

@FELIPE CABRAL

Your /etc/hosts should look like this in the below example I used 192.168.0.124 as the IP of edcserver.domain please use the appropriate IP

The first 2 lines shouldn't be tampered with

 # cat /etc/hosts
127.0.0.1 localhost localhost.localdomain
::1 localhost localhost.localdomain
192.168.0.124  edcserver.domain  edcserver

Now to overcome the error 1 other registered host (screenshot 2018-05-31-18-09-36-window.png) you have to delete this entry from 2 tables hoststate and hosts in the Ambari database.

I have attached the procedure below, the assumption is as follows

  • Ambari database name = ambari
  • Ambari database type= MySQL
  • Ambari database password =Secret1234

If you are using postgres please adapt accordingly the SQL

Log on the DB

# mysql -u root -pSecret1234 

some output here

Check the databases

mysql> show databases; 
+--------------------+ 
| Database           | 
+--------------------+ 
| information_schema | 
| ambari             | 
| hive               | 
| mysql              | 
| oozie              | 
| ranger             | 
| rangerkms          | 
| test               | 
+--------------------+ 
8 rows in set (0.00 sec)

Connect to the Ambari database

mysql> use ambari; 
Reading table information for completion of table and column names 
You can turn off this feature to get a quicker startup with -A 
Database changed 

Check the structure if hoststate table we need to delete the host_id

mysql> describe hoststate; 
+-------------------+--------------+------+-----+---------+-------+ 
| Field | Type | Null | Key | Default | Extra | 
+-------------------+--------------+------+-----+---------+-------+ 
| agent_version | varchar(255) | NO | | NULL | | 
| available_mem | bigint(20) | NO | | NULL | | 
| current_state     | varchar(255) | NO   |     | NULL    |       | 
| health_status     | varchar(255) | YES  |     | NULL    |       | 
| host_id           | bigint(20)   | NO   | PRI | NULL    |       | 
| time_in_state     | bigint(20)   | NO   |     | NULL    |       | 
| maintenance_state | varchar(512) | YES  |     | NULL    |       | 
+-------------------+--------------+------+-----+---------+-------+ 
7 rows in set (0.03 sec) 
mysql> select host_id from hoststate; 
+---------+ 
| host_id | 
+---------+ 
| 1       | 
+---------+ 
1 row in set (0.04 sec) 
mysql> Delete from hoststate where host_id=1; 
Query OK, 1 rows affected (0.00 sec) 

Check the hosts table

mysql> describe hosts; 
+------------------------+---------------+------+-----+---------+-------+ 
| Field 		 | Type 	 | Null | Key | Default | Extra | 
+------------------------+---------------+------+-----+---------+-------+ 
| host_id 		 | bigint(20) 	 | NO 	| PRI | NULL 	| 	| 
| host_name 		 | varchar(255)  | NO 	| UNI | NULL 	| 	| 
| cpu_count 		 | int(11) 	 | NO 	|     | NULL    |       | 
| cpu_info  		 | varchar(255)  | NO   |     | NULL    |       | 
| discovery_status 	 | varchar(2000) | NO   |     | NULL    |       | 
| host_attributes 	 | longtext 	 | NO   |     | NULL    |       | 
| ipv4 			 | varchar(255)  | YES  |     | NULL    |       | 
| ipv6 			 | varchar(255)  | YES  |     | NULL    |       | 
| last_registration_time | bigint(20)    | NO   |     | NULL    |       | 
| os_arch 		 | varchar(255)  | NO   |     | NULL    |       | 
| os_info 		 | varchar(1000) | NO   |     | NULL    |       | 
| os_type 		 | varchar(255)  | NO   |     | NULL    |       | 
| ph_cpu_count 		 | int(11)       | YES  |     | NULL    |       | 
| public_host_name 	 | varchar(255)  | YES  |     | NULL    |       | 
| rack_info 		 | varchar(255)  | NO   |     | NULL    |       | 
| total_mem 		 | bigint(20)    | NO   |     | NULL    |       | 
+------------------------+---------------+------+-----+---------+-------+ 
16 rows in set (0.02 sec) 

Retrieve and delete the host with the matching id 1

mysql> select host_id from hosts; 
+---------+ 
| host_id | 
+---------+ 
| 1 	  | 
+---------+ 
1 row in set (0.00 sec) 
mysql> Delete from hoststate where host_id=1; 
Query OK, 1 rows affected (0.00 sec)

And restart ambari-server and proceed with the registration all should be fine.

Please revert If you found this answer addressed your question, please take a moment to log in and click the "accept" link on the answer.

View solution in original post

Re: Error add node into cluster

New Contributor

Nice! now it works fine.

I got passed of installation part. But now, error message are showed when i try configure the service into te machine, follow.

I have create the directory and give all permission to them, but not works.

greenshot-2018-06-01-120718.png

Other erros that are showed it just insert password to admin ou root database. I did it. But the HDFS and YARN can't do it.

Guy! It's works. The problem above question was the directory that can't initiate with '/home'!

I change this part and the installation continue.

Thanks so much!!! :)

Don't have an account?
Coming from Hortonworks? Activate your account here