Member since
11-24-2015
223
Posts
10
Kudos Received
0
Solutions
04-04-2017
06:38 PM
You're probably running in the leftover tombstone from a delete: https://hbase.apache.org/book.html#_delete Compact your table and then run the import.
... View more
04-03-2017
01:12 PM
Thanks for the response. In Ambari -> Zookeeper : I can see three entries for zookeeper server and 2 zookeeper clients installed. Not sure about server ip mapping and myid match - can you give more detail pls? [root@hdtesting1 etc]# netstat -tlpn | grep 2181
tcp 0 0 0.0.0.0:2181 0.0.0.0:* LISTEN 11297/java zoo.cfg : maxClientCnxns=50
# The number of milliseconds of each tick
tickTime=2000
# The number of ticks that the initial
# synchronization phase can take
initLimit=10
# The number of ticks that can pass between
# sending a request and getting an acknowledgement
syncLimit=5
# the directory where the snapshot is stored.
dataDir=/var/lib/zookeeper
# the port at which the clients will connect
clientPort=2181
... View more
08-14-2018
09:50 AM
Is there a way to ensure that after the migration using mysqldump partitions can also be moved to the new metastore. I have tried moving metastore from one environement to other but i dont see partitions and stats , any idea how can i address that?
... View more
03-21-2017
10:35 PM
1 Kudo
Below is a good read on Determining HDP Memory configs: http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.4/bk_installing_manually_book/content/determine-hdp-memory-config.html
... View more
03-07-2017
07:14 PM
See if these docs help: http://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-views/content/ch_using_falcon_view.html and http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.3/bk_data-movement-and-integration/content/index.html.
... View more
02-07-2017
03:43 PM
It is point in time (time at which distcp runs). It can be automated using scripts. It is replication because you are replicating data. you are confusing replication with real time replication. Replication doesn't have to be real time. And it is impossible because of physics that any change in data in one cluster (or anyother database) will be reflected in the other. I'll give you the example of my HBase use case. We were using Active-Active replication. But even then, we knew that there might be a situation where data will be written to one data center and a power failure will occur as data is being replicated to the remote data center and some data will not be replicated (let's say up to 10 to seconds of data). The only other way to make sure that this does not happen is to "Ack" only when data has been written to the remote data center. This slows down every write and we had 10's of thousands of writes per second. See, you have to make a choice. If you would like to have 100 percent sync then you have to make sure that you ack every single record being written slowing down all your writes. Or you can do asynchronous replication which works 99.99% of the time but in case of network issues between two data centers, you know some data some time will not be replicated right away. There is absolutely nothing technology can do anything here. This is simple Physics.
... View more
02-06-2019
05:31 AM
can any one please send me the ambari installation document on ubuntu
... View more
- « Previous
- Next »