Member since
10-06-2015
45
Posts
54
Kudos Received
0
Solutions
01-11-2016
07:21 PM
3 Kudos
I have a requirement to periodically restart all cluster nodes at the machine level. Assume I've done an FSCK before starting to confirm that all blocks are fully replicated. Question is, as I restart each node in turn, will the NameNode notice that any block on that node is under-replicated and put those blocks on the replication queue? If this does happen, will it automatically remove those blocks when the data node comes back online and reports it's blocks to the NN? Note, this is a hardware restart, so the Ambari rolling restart doesn't do the job.
... View more
Labels:
- Labels:
-
Apache Hadoop
01-06-2016
07:37 PM
3 Kudos
Within a cluster we have no trouble executing commands agains an HA NameNode using the NameServiceID. But it doesn't work when doing discp from one cluster to another because the clusters are unaware of each other's mapping of nodes to NameServiceID. How does one do this?
... View more
Labels:
- Labels:
-
Apache Hadoop
01-04-2016
05:24 PM
4 Kudos
We have two use cases--one is the normal slight imbalance that can creep up gradually and the other is when we add new nodes. Ten new nodes can be 100TB+ to move around--it can take a very long time with normal dfs.network.bandwidth.persecond setting. What's a good strategy? Is it reasonable to use chron to reset the value during off hours? What's the best practice? Also, does rebalancing defer to normal processing dynamically?
... View more
Labels:
- Labels:
-
Apache Hadoop
12-09-2015
04:47 PM
The documentation seems to suggest that the normal mode of use would be to have one reconstituted replica sitting around and that reconstituting an encoded block would be done only if this isn't the case. Keeping a block by default would eliminate most of the space savings because the data would expand from 1.6 to 2.6 times the raw file size. Why not have a policy that for leaves a single size copy for a limited time after a block is used? A "working set" as it were, so if you've used a block in the last X hours the decoded block won't be deleted.
... View more
12-08-2015
10:29 PM
1 Kudo
The admins want to know why every service has its own account ID, and is there any harm is using the same account for all? The cluster will be tightly secured. What is the best practice?
... View more
12-06-2015
12:03 AM
1 Kudo
Hadoop has long stressed moving the code to the data, both because it's faster to move the code than to move the data, and more importantly because the network is a limited shared resource that can easily be swamped. Erasure coding would seem to require that a large proportion of the data must move across the network because the contents of a single block will reside on multiple nodes. This would presumably apply not just the ToR switch, but the shared network as well, if the ability to tolerate the loss of a rack is preserved. Is this true and how are these principles reconciled?
... View more
Labels:
- Labels:
-
Apache Hadoop
11-23-2015
06:10 PM
1 Kudo
Your inode article is a great addition to David's answer. I'm puzzled though that any machine would run out of inodes before running out of disk space---it would require a strange configuration of the file system, wouldn't it? Was someone trying to save on inode allocation by assuming the average file would be larger? I can't think of any other reason to stray from the defaults. Any idea why?
... View more
11-20-2015
08:26 PM
Thanks Ancil. I'm still curious about what can be done from inside Hadoop. The federation of queries is particularly interesting becauese you don't always want to import the data into HDFS.
... View more
11-20-2015
03:44 PM
1 Kudo
Thanks for the reply. Yes, I read that page--the problem is trying to confirm whether the version of the connector in that tarball that this leads to, which seems to be for HDP 2.3 works with 2.2.4. Can't seem to locate one specifically for 2.2.4.
... View more