08-19-2014 10:05 AM
Thanks for your help Romain.
The sharelib is the one used for Yarn: oozie-sharelib-yarn.tar.gz
I've enclosed the configuration of the job from Oozie, but this looks like it is using Yarn. The job starts, but never finishes, instead it repeats Heart beat over and over. I see an entry in the log that refers to port 50030, which is why it looks like it is using MRv1. But I can see the job in Yarn's ResourceManager, it is RUNNING, but never finishes until killed.
08-19-2014 11:23 AM
08-19-2014 12:52 PM
08-19-2014 01:59 PM
08-19-2014 02:34 PM
08-19-2014 05:55 PM
08-20-2014 09:29 AM
Thank you so much for your help, and for sticking with me through this problem. I have resolved the issue. There were actually two problems. After the upgrade to CDH 5, I had to stop Oozie and Install Sharelib. Finally, in YARN I had to adjust the resources. The Java Heap Size had been set to 50 MB when 8 GB of memory is available to the node (I set heap memory to 1 GB on the nodes and resource manager). I don't know why the CDH update would default to such a low number - this made YARN completely unusable. This explains why jobs would hang forever as there was not enough resources available. However, the logs did indicate this problem.
I have one last question, how much memory do you give to the Java heap on the resource manager, under Java Heap Size of ResourceManager in Bytes, when the nodes are given 1 GB. I gave this 1 GB to resolve the problem, but I'm not sure if that is enough. And what about the Container sizes?
12-18-2014 01:04 AM
I met same issue as yours.
I run a simple pig such as Load 'File'; Dump data;
However, the pig cannot be completed and the logs always should 0% complete.
I search the log and find the map progress is always running but not complete. I have adjusted the values of settings you list before. But it looks like the issue still exist.
Do you have other ideas on it?
Thanks very much
12-18-2014 09:09 AM
When I've seen Pig scripts show 0% complete and never finish, I've usually resolved this by adjusting Yarn. How many nodes are you running in your cluster? How much memory is available to your nodes?