Created on 11-03-2016 08:43 AM - edited 09-16-2022 03:46 AM
So far I have a single node cluster (cluster-1 with host-a) . Now, I am trying to add a new host which is an AWS EC2 instance.
I came across this article:
http://hortonworks.com/blog/deploying-hadoop-cluster-amazon-ec2-hortonworks/
It says that
Remember the list of private DNS names that you had copied down to a text file. We will pull out the list and paste it in the Target host input box. We will also upload the private key that we have been using on this page:
Then we also need to open up the ports for IPs internal to the datacenter:
The IP of the server I am trying to add as a new host (let's call it host-B) is not internal to the data center where my existing single node cluster is hosted.
What all I do I need to do for adding this new host that is not internal to the data-center where my existing single node cluster is hosted?
Amazon server gives:
a hostname
a public DNS
a private DNS
a Public IP
a private IP
What exactly do I need to add in mt /etc/hosts of the existing host (host-A) in the cluster for it to be able to access the new host (host-B) that is hosted with AWS.
Also, what needs to go in /etc/hosts of the new host (host-b)
Please suggest!
My problem is that even though the host was added, it is not in the list of live hosts. Also, heartbeat is available
Created 11-05-2016 01:24 AM
Hadoop clusters spanning multiple data centers is not supported and can lead to unsatisfactory results.
Created 11-05-2016 01:24 AM
Hadoop clusters spanning multiple data centers is not supported and can lead to unsatisfactory results.
Created 11-05-2016 02:10 AM
You cannot do this. This is not supported as @Artem Ervits has already stated. Imagine what would happen to writes when clusters span multiple data centers? Remember, networks are assumed to be unreliable and unsecured.
Now, I hate to say this and please don't do it as it is unsupported but amazon offers VPC which makes AWS an extension of your network using a VPN.