Support Questions

Find answers, ask questions, and share your expertise

Resource manager is not working and process exiting unexpeted. Memeory issue

avatar
Explorer

Hi All,

 

I have a RHEL 7.5 AWS M5.xlarge instance. It is a 16GB, 4 core machine.

I have done the path B manual installation of CDH using CDH manager, the installation was successfull.

 

After, installation the resource manager in yarn is down, i have tried some memory settings but it is not working. I am continuesly facing yarn tuning issues, it works for couple daya and goes down.

 

Can someone help me with following memeory allocation for a 16GB single node cluster:

 

yarn.app.mapreduce.am.resource.mb

mapreduce.map.memory.mb

mapreduce.reduce.memory.mb

mapreduce.job.heap.memory-mb.ratio

Client Java Heap Size in Bytes

Java Heap Size of JobHistory Server in Bytes

memory.soft_limit_in_bytes

Java Heap Size of NodeManager in Bytes

Container Memory - yarn.nodemanager.resource.memory-mb
Java Heap Size of ResourceManager in Bytes
yarn.scheduler.minimum-allocation-mb
yarn.scheduler.increment-allocation-mb
yarn.scheduler.maximum-allocation-mb
 
Its a single cluster machine.
 
Thanks in advance.

 

12 REPLIES 12

avatar
Explorer

I have not run any job yet, this error comes even before running a job. When it tries to invoke the internal dr.who job, it goes down.

 

I have done the memory setting as per your suggestion earlier.

 

Are there any other paramters i should change. Following is my current settings:

 

yarn.app.mapreduce.am.resource.mb - 2 gb

mapreduce.map.memory.mb - 1 gb

mapreduce.reduce.memory.mb -   1 gb

mapreduce.job.heap.memory-mb.ratio - 0.8

Client Java Heap Size in Bytes - 787 mb

Java Heap Size of JobHistory Server in Bytes - 431 mb

memory.soft_limit_in_bytes - -1 mb

memory.hard_limit_in_bytes - -1 mb

Java Heap Size of NodeManager in Bytes - 1 gb

Container Memory - yarn.nodemanager.resource.memory-mb - 2gb
Java Heap Size of ResourceManager in Bytes - 1 gb
yarn.scheduler.minimum-allocation-mb - 1 gb
yarn.scheduler.increment-allocation-mb - 512 mb
yarn.scheduler.maximum-allocation-mb - 2 gb

avatar
Master Guru

@Riteshk,

 

In the logs you provided, there are links to view the application logs.  The error 143 return code does not mean there was a memory problem for sure.  In order to tell what happened to the containers, you'll need to look at the logs more closely for that job.

avatar
Explorer

The Resource manager starts and shuts down in couple of mins. It shows unpexected exits and there are no error in logs for resource manager.

 

Is RHEL 7.5 m5.xlarge (4CPU, 16GB) machine capable of running cloudera with spark2,oozie,yarn,hue,hive and

also cloudera manager?

 

Am i missing something here? Please help.