Member since
06-25-2018
10
Posts
1
Kudos Received
0
Solutions
04-13-2019
01:40 AM
Harsh J: Thanks for the help on the previous issue. We finally resolved the issue. It was due to an undocumented port required in the CDH 6.2 to CDH 6.2 distcp. Now, we are migrating the task over to Oozie and having some trouble. Could you elaborate a bit more or give us some links or pointers? Thanks. We could not find "mapreduce.job.hdfs-servers" . Where is that?
... View more
04-10-2019
12:18 AM
Hi. I am having a "What the heck" moment. Could someone please explain the theory behind this. I have always presumed that Sqoop - unlike other MR processes that might require the entire dataset to be in memory to work - should not ever have a OOM issue. Afterall, it is using its memory as a buffer, copying the data from DB to the staging area in HDFS, and when complete, moving from staging to --target-dir.
So, we were moving a fairly large DB (500GB) but our client would only allow us to use 1 mapper (don't ask why...gulp). About 90 minutes into the process, it terminated with:
Container is running beyond the 'PHYSICAL' memory limit. Current usage: 1.0Gib of 1 GB physical memory used; 2.7GB of 2.1 GB virtual memory used. Killing container
This is really confusing me. I suppose I can solve the problem by
a) increasing the vmem/pmem ratio (yarn.nodemanager.vmem-pmem-ratio = xyz)
OR
b) not checking for this error (yarn.nodemanager.vmem-check-enabled = false).
But WHY is this error coming up?
Thanks in advance and cheers.
... View more
Labels:
- Labels:
-
Apache Sqoop
-
Apache YARN
04-04-2019
04:58 AM
Thank you. Because of the client’s security measures, we are unable to disperse the log files generated. This, of course, makes everything so much more difficult.
... View more
04-02-2019
06:24 AM
Thanks for your reply. Could we ask a related question? Our client is very reluctant to open the ports on these 2 clusters. Could you tell us what ports need to be open for distcp to function properly? After many fails, our client has briefly allowed all the ports to be open. With that change, distcp if working properly. We have already looked at the ports specified in https://www.cloudera.com/documentation/enterprise/latest/topics/install_ports_distcp.html#topic_9_1. Are there any hidden ports or secondary ports beyond the above documentation that could be causing the problem?
... View more
03-31-2019
06:26 AM
1 Kudo
Hi. We have a client who has 2 clusters. On the security cluster, they have sensitive data that they redact and copy to the analysis cluster. For security reasons, they would like to minimize the number of open ports on the security cluster. We have successfully tested using distcp from the shell to copy the data with port 8020 open. They would now like to automate the process through oozie. In testing, we have run into an error that port 8042 (Node Manager External Port) is not open.
We do not understand why distcp works fine without port 8042 available when run through the shell but fails when called through Oozie.
Any help would be appreciated. Thanks.
Henry
... View more
Labels:
- Labels:
-
Apache Oozie
-
Security
01-09-2019
04:38 PM
Thanks for the response. A quick reaction…. Why not 5 master and 5 tservers, then? Might as well put a master on all of them was my gut reaction. What would be the disadvantage of that given that it is light weight? If I were to answer my own question, I might say that since there is no automatic process to replace a faulty master, 5 masters doesn’t really give that much peace of mind anyway.
... View more
01-08-2019
04:41 PM
Thanks for your quick response. Could you give me a clarification. Am I to understand that there is no automatic recovery mechanism for Masters? So in all installations, if a Master fails, we have to always manually recover using the steps you provided in your link?
... View more
01-07-2019
10:44 PM
We have a client who wants to setup a small separate special purpose cluster. It will consist of 5 nodes and they want to run Kudu on it. If we configure with 3 Master nodes and 3 Tablet Servers, at least one of the nodes will have both Master roles as well as Tablet Server role. In any case, in this scenario, if hypothetically the node with both roles fail, and the replication is set to 3, neither the master tablet nor any of the table tablets will be able to satisfy the replication factor since there are now only 2 master roles and 2 table servers available. Since this is still within the (n-1)/2 failures, will the service still run? After 5 minutes, and when the nodes are determined to be permanent failures, a new follower needs to be created but since there are only 2 tablet servers still functioning, what will happen? Will the system limp along, still functioning or will it fail? Thanks in advance for any help.
... View more
Labels:
- Labels:
-
Apache Impala
-
Apache Kudu