Member since
02-27-2017
171
Posts
9
Kudos Received
0
Solutions
02-14-2017
05:24 AM
@Attila Kanto
I got to know yesterday that cloudbreak cannot install hadoop clusters into already running VM's. is that correct? you mentioned that we donot need to provide fqdn in blueprint file. Rather it should be present in cluster creation file. But in cloudbreak UI i donot see any way to upload cluster creation template? Could you please suggest. We are stuck as we already have 4 VM's running in azure but we donot know how to use cloudbreak to install hadoop in those VM's? Much appreciated. Thanks Rahul
... View more
02-13-2017
04:20 PM
Hi, I have found the solution to above mentioned problem. Actually i have changed the cidr to 0.0.0.0/0 hile creating network security group and it worked. But the point is that it has started 4 new instances in azure rather than deploying it in 4 existing VM. I already have 4 VM's running and i want to deploy these services in those 4 VM's. How to make it work if i want to give CIDR for my already existing network? Thanks Rahul
... View more
02-13-2017
03:32 PM
@rkovacs Any help would be appreciated? Is it due to the fact that i am using latest version of cloudbreak 1.6.2? Thanks Rahul
... View more
02-13-2017
03:26 PM
@rdoktorics Any help on this would be appreciated? Thanks
... View more
02-13-2017
03:24 PM
I am deploying 4 node cluster on azure using cloudbreak. It seems like it got stuck in between as it it not moving forward after creating infrastructure(stuck at "Infrastructure metadat collection finished". Kindly see the logs mentioned below 2/13/2017 8:31:06 PM shaneconew - create in progress: Creating infrastructure2/13/2017 8:36:13 PM shaneconew - update in progress: Infrastructure creation took 307 seconds2/13/2017 8:36:16 PM shaneconew - update in progress: Infrastructure metadata collection finished My cloudbreak version if 1.6.2(latest one) and if i check logs using "tail -100f cbreak-20170210-082017.log" i can see connection timed out. Is the deployment running at the back or not? /cbreak_cloudbreak_1 | 2017-02-13 15:23:50,259 [reactorDispatcher-31] checkStatus:37 INFO c.s.c.s.s.f.SshCheckerTask - [owner:b87a672d-bb93-4a0b-843d-4b8c15f1968d] [type:STACK] [id:17] [name:shaneconew] Failed to connect ssh: Connection timed out (Connection timed out)
/cbreak_cloudbreak_1 | 2017-02-13 15:23:55,276 [reactorDispatcher-31] pollWithTimeout:30 INFO c.s.c.s.PollingService - [owner:b87a672d-bb93-4a0b-843d-4b8c15f1968d] [type:STACK] [id:17] [name:shaneconew] Polling attempt 8.
... View more
Labels:
- Labels:
-
Hortonworks Cloudbreak
02-06-2017
11:37 AM
Hi Peter, Just a small question. My Spark oozie workflows keeps on running from long time. When i checked oozie logs i found it is trying to connect to port 8032 on sandbox.hortonworks.com. I donot know why it is going to 8032 instead of 8050 although i have mentioned 8050 in my job.properties. Any idea? Thanks Rahul
... View more
02-06-2017
09:22 AM
Hi Peter, Thank you so much for such a clear answer. I tried the steps mentioned below and set the individual values as 4096 and 3072 but my job failed due to "MAP capability required is more than the supported max container capability in the cluster". I checked the properties "mapreduce.map.memory.mb" and "mapreduce.map.java.opts" in mapred-site.xml and their values mentioned are 250 and -Xmx200m. So This might be the reason my job is getting killed as it is requesting container size more than default values. Any workaround for this? If i update the values in mapred-site.xml to above mentioned values, then which services i need to restart to reflect those changes? Or can it be resolved in any other way? By the way I am running HDP 2.5 Thanks Rahul
... View more
02-06-2017
08:14 AM
Hi Peter, I have already specified my driver-memory to 5g in spark_opts in workflow.xml. I am still getting the same error. Does this has something to do with memory assigned to HDP 2.5 on Virtual box. In my case it is 5 GB?
... View more
02-06-2017
08:14 AM
Hi Peter, I have already specified my driver-memory to 5g in spark_opts in workflow.xml. I am still getting the same error. Does this has something to do with memory assigned to HDP 2.5 on Virtual box. In my case it is 5 GB?
... View more
02-06-2017
08:01 AM
I am trying to run simple spark job using oozie workflow scheduler. And I am getting error as System memory 202375168 must be at least 4.718592E8. Please use a larger heap size. i have assigned 5 GB to my HDP sandbox on Virtual box. I have created a spark jar on my local machine and uploaded the jar to HDP sandbox. My workflow.xml looks like as below. <workflow-app name="samplespark-wf" xmlns="uri:oozie:workflow:0.4"> <start to="sparkjob"/>
<action name="sparkjob">
<spark xmlns="uri:oozie:spark-action:0.1"> <job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node> <master>local[1]</master>
<name>Spark Test</name>
<class>main.scala.RDDscala.RDD1</class> <jar>${nameNode}/spark_oozie_action/sparkrdd_2.11-0.0.1.jar</jar>
<spark-opts>--driver-memory 5g --num-executors 1</spark-opts> </spark>
<ok to="end"/>
<error to="fail"/> </action>
<kill name="fail">
<message>Shell action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<kill name="fail-output">
<message>Incorrect output, expected [Hello Oozie] but was [${wf:actionData('shell-node')['my_output']}]</message>
</kill>
<end name="end"/>
</workflow-app> This program runs fine on local system command prompt with below mentioned command spark-submit --class main.scala.RDDscala.RDD2 --master local target\scala-2.11\sparkrdd_2.11-0.0.1.jar Any help would be appreciated
... View more
Labels:
- Labels:
-
Apache Oozie