Member since
09-24-2015
178
Posts
113
Kudos Received
28
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3493 | 05-25-2016 02:39 AM | |
3703 | 05-03-2016 01:27 PM | |
871 | 04-26-2016 07:59 PM | |
14716 | 03-24-2016 04:10 PM | |
2167 | 02-02-2016 11:50 PM |
11-17-2015
11:51 PM
HDFS restart is required after the change for the new config to take effect permanently. For changing the params as part of the job, this should work - export HADOOP_OPTS="-Dipc.server.tcpnodelay=true"
... View more
11-17-2015
06:19 PM
1 Kudo
Is there a way to configure a non-default replication factor for a HDFS directory such that all future files and sub-directories in that directory use that specific replication factor? Currently, we are using a work around of running a daemon process to set the replication factor for all files in the required directory. Is there a better way to do this? while true; do
hdfs dfs -setrep -w 2 /tmp/
sleep 30
done I see at one point there was this JIRA https://issues.apache.org/jira/browse/HDFS-199 opened but is blocked by this JIRA [https://issues.apache.org/jira/browse/HADOOP-4771]
... View more
Labels:
- Labels:
-
Apache Hadoop
11-17-2015
05:20 PM
I don't think this feature is supported natively by Oozie or Falcon. It is a great feature request though. Since Falcon already has the knowledge of DR cluster, it should just replicate the jobs in the secondary cluster. I think the other part of the requirements is that admin should be able to turn on all of the jobs on DR cluster using a single flag / step.
... View more
11-17-2015
04:08 PM
Is there a recommended rules engine for Hadoop? Has anyone tested Drool with Hive? The idea is to have a repository of rules, and the engine should read the rules and apply them to the data. These rules could be simple (e.g. value >10) or complex (e.g. average age is 40% greater than the standard deviation of the population). I know that I can write these rules manually in HIVE, but we would like to let the end users make changes on the rules without having us re-write the rule.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
11-02-2015
03:39 PM
2 Kudos
I know CB creates docker containers with HDP components to provision HDP cluster but can CB also work with a cloud infrastructure that uses Docker as Virtualization layer?
... View more
Labels:
- Labels:
-
Docker
-
Hortonworks Cloudbreak
10-27-2015
03:29 AM
1 Kudo
Change the yarn-log4j configuration as shown. Make sure that the properties are uncommented.
... View more
10-26-2015
01:31 PM
When a new Configuration group is created in Ambari and a newly added host is added to that group, I am unable to edit the settings / properties / configurations for that group. Everything is greyed out. Verified that admin user is logged in. Also tried creating a brand new configuration group as well as duplicating an existing configuration group but no luck. What else needs to be done to be able to edit the properties? I am specifically looking to change the namenode and datanode directories for HDFS. The existing nodes have /grid/[0-3] but the new node only has 1 disk thats mounted to /grid/0 so no [1-3].
... View more
Labels:
- Labels:
-
Apache Ambari
10-26-2015
03:46 AM
1 Kudo
It is not possible to tell the root cause of the issue by looking at the trace you provided. The log tells that the MR step failed but does not say why. It is possible to get the exact details of the failure though, using the steps below - 1) Click on the job instance in Oozie 2) On the next page, double click on the step that failed. 3) On popup click on the small lens icon button to pull up the log. 4) On the job appilcation page on YARN UI, click on logs and get more information. In my experience, the error messages are very direct on this page and will tell exactly what the problem is.
... View more
10-26-2015
03:27 AM
It could be because the user home directory does not exist on HDFS.
... View more