Member since
07-05-2016
25
Posts
45
Kudos Received
3
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1751 | 12-01-2017 01:25 AM | |
13449 | 11-27-2017 10:51 PM | |
1202 | 03-13-2017 08:45 PM |
12-01-2017
01:25 AM
2 Kudos
What is the conf dfs.datanode.data.dir set to? I suspect it is set to different directories under the same physical disk.
... View more
11-28-2017
12:25 AM
1 Kudo
For recoverLease using CLI, see https://community.hortonworks.com/questions/146012/force-closing-a-hdfs-file-still-open-because-uncor.html?childToView=146021#answer-146021
... View more
11-27-2017
10:51 PM
3 Kudos
"Sleep and retry" is good way to handle the "not have enough number of replicas" problem. For the "already the current lease holder" problem, you may call DistributedFileSystem.recoverLease(Path) to force lease recovery. Hope it helps.
... View more
11-27-2017
10:22 PM
Suppose my user name is "nicholas" and "nicholas" is already configured as a proxy user. Now, is there a way for "nicholas" to run a dfs command (say mkdir) as another user "foo"?
... View more
Labels:
- Labels:
-
Apache Hadoop
05-09-2017
11:20 PM
> Is there any maximum configurable value of ipc.maximum.data.length? Hadoop does not enforce a maximum. > Can we change this value above 128MB? Yes, you may change it to 192MB or 256MB to get around the current issue.
... View more
03-13-2017
08:45 PM
1 Kudo
No, it is a non-disruptive procedure, provided that the cluster is healthy and is not under a heavy load. One of reasons to do so is for upgrading namenode, either software or hardware. During a namenode failover, the jobs and clients application will be redirected from the old active namenode to the new active namenode. Of course, they have to wait until the new active namenode becomes ready so that they are slowed down. In this sense, we are better to perform the failover operation when the cluster is idea or under a small load. Hope it helps.
... View more
03-13-2017
06:55 PM
What is your version of Hadoop? Could you post the output from "hadoop -version"?
... View more
03-13-2017
06:49 PM
Yes, the audit log will serve the purpose. Note that, in some cases,
it is not straightforward to search the log for deletion since a
directory (or a file) may not be deleted directly -- it may be deleted
as a part of the deletion of its parent/ancestor directory. So we
should first search the full path in the log. If it is not found,
search the parent directory path and so on. It will be more complicated if deletion and re-creation occurred repeatedly. For example 1) user A: create /foo 2) user A: create /foo/bar 3) user A: del /foo 4) user B: create /foo 5) user B: del /foo
Who has deleted /foo/bar? It is easy to mistakenly take user B as the answer. B is the last user deleted foo but B
is not the user deleted /foo/bar. In such case, we should first
determine when the target directory/file is created and then search what
happened of it starting from the creation time. You can imagine that it is even harder to find out the correct answer if the path or the parent/ancestor paths are moved/renamed. We need to pay extra attention if the rename operation is involved.
... View more