Created 08-18-2017 04:11 AM
Hello All,
I noticed agent status bad notification for running 3-node cluster in cloudera manager CDH5.4 console.
When I tried to access the link for agent log, it gave me error.
I checked logs on the corresponding server at /var/log/cloudera-scm-agent , there I fiound out below error.
TP/1.1" 200 2120 "" "Java/1.7.0_67"
[08/Aug/2017 05:22:47 +0000] 8060 MainThread agent INFO Process with same id has changed: 2940-host-inspector.
[08/Aug/2017 05:22:47 +0000] 8060 MainThread agent INFO Deactivating process 2940-host-inspector
[08/Aug/2017 05:22:48 +0000] 8060 Metadata-Plugin navigator_plugin INFO stopping Metadata Plugin for host-inspector with pipelines []
[08/Aug/2017 05:22:48 +0000] 8060 Metadata-Plugin navigator_plugin_pipeline INFO Stopping Navigator Plugin Pipeline '' for host-inspector (log dir: None)
[08/Aug/2017 05:22:49 +0000] 8060 Audit-Plugin navigator_plugin INFO stopping Audit Plugin for host-inspector with pipelines []
[08/Aug/2017 05:22:49 +0000] 8060 Audit-Plugin navigator_plugin_pipeline INFO Stopping Navigator Plugin Pipeline '' for host-inspector (log dir: None)
[08/Aug/2017 05:26:00 +0000] 8060 MonitorDaemon-Reporter throttling_logger ERROR (1 skipped) Error sending messages to firehose: mgmt-SERVICEMONITOR-0fc5431c91f288535e98f1ed6d2d7836
Traceback (most recent call last):
File "/usr/lib64/cmf/agent/src/cmf/monitor/firehose.py", line 74, in _send
self._requestor.request('sendAgentMessages', dict(messages=messages))
File "/usr/lib64/cmf/agent/build/env/lib/python2.6/site-packages/avro-1.6.3-py2.6.egg/avro/ipc.py", line 139, in request
return self.issue_request(call_request, message_name, request_datum)
File "/usr/lib64/cmf/agent/build/env/lib/python2.6/site-packages/avro-1.6.3-py2.6.egg/avro/ipc.py", line 249, in issue_request
call_response = self.transceiver.transceive(call_request)
File "/usr/lib64/cmf/agent/build/env/lib/python2.6/site-packages/avro-1.6.3-py2.6.egg/avro/ipc.py", line 478, in transceive
result = self.read_framed_message()
File "/usr/lib64/cmf/agent/build/env/lib/python2.6/site-packages/avro-1.6.3-py2.6.egg/avro/ipc.py", line 482, in read_framed_message
response = self.conn.getresponse()
File "/usr/lib64/python2.6/httplib.py", line 990, in getresponse
response.begin()
File "/usr/lib64/python2.6/httplib.py", line 391, in begin
version, status, reason = self._read_status()
File "/usr/lib64/python2.6/httplib.py", line 349, in _read_status
line = self.fp.readline()
File "/usr/lib64/python2.6/socket.py", line 433, in readline
data = recv(1)
error: [Errno 104] Connection reset by peer
Below are the outputs that I checked.
[root@LinuxUL cloudera-scm-agent]# cat /etc/hosts
127.0.0.1 localhost
10.68.200.34 LinuxUL.ad.infosys.com LinuxUL
10.68.200.152 linux152.ad.infosys.com linux152
10.68.200.170 linux170.ad.infosys.com linux170
172.21.5.224 nfrsat01.ad.infosys.com nfrsat01
10.67.200.77 blrsat06.ad.infosys.com blrsat06
[root@LinuxUL ~]# hostname -f
LinuxUL.ad.infosys.com
[root@LinuxUL ~]# python -c 'import socket; print socket.getfqdn(), socket.gethostbyname(socket.getfqdn())'
LinuxUL.ad.infosys.com 10.68.200.34
Also, selinux and iptables are disabled.
Please suggest.
Thanks,
Priya
Created 08-22-2017 03:04 AM
Hi cdhhadoop,
Cloudera agent is completely down? It happens in more servers?
Can you provide /var/log/cloudera-scm-agent/cloudera-scm-agent.out output?
Can you provide the output of the next commands?:
$ netstat -ltnp | grep :9000 $ source /etc/cloudera-scm-agent/config.ini &>/dev/null
$ ping -w1 $server_host $ telnet $server_host $server_port
Regards,
Marc Casajus
Created 08-22-2017 03:20 AM
Created 08-22-2017 04:00 AM
Please run:
$ source /etc/cloudera-scm-agent/config.ini &>/dev/null $ telnet $server_host $server_port
&> without any space between & and >.
Regards,
Marc.
Created 08-23-2017 04:29 AM
Created 08-23-2017 04:33 AM
Created 08-23-2017 10:20 PM
Created 08-28-2017 03:40 AM
Created 01-31-2019 10:31 PM
Hi Team ,
can you please help on below error.
connection reset by peer error.
Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead.
19/01/31 18:43:18 INFO spark.SparkContext: Running Spark version 2.2.0.cloudera1
19/01/31 18:43:19 WARN spark.SparkConf: In Spark 1.0 and later spark.local.dir will be overridden by the value set by the cluster manager (via SPARK_LOCAL_DIRS in mesos/standalone and LOCAL_DIRS in YARN).
19/01/31 18:43:19 INFO spark.SparkContext: Submitted application: Prime_CEP_BGFR_1309_Process Rates/Other Errors
19/01/31 18:43:19 INFO spark.SecurityManager: Changing view acls to: ggbmgphdpngrp
19/01/31 18:43:19 INFO spark.SecurityManager: Changing modify acls to: ggbmgphdpngrp
19/01/31 18:43:19 INFO spark.SecurityManager: Changing view acls groups to:
19/01/31 18:43:19 INFO spark.SecurityManager: Changing modify acls groups to:
19/01/31 18:43:19 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ggbmgphdpngrp); groups with view permissions: Set(); users with modify permissions: Set(ggbmgphdpngrp); groups with modify permissions: Set()
19/01/31 18:43:19 INFO util.Utils: Successfully started service 'sparkDriver' on port 50000.
19/01/31 18:43:19 INFO spark.SparkEnv: Registering MapOutputTracker
19/01/31 18:43:19 INFO spark.SparkEnv: Registering BlockManagerMaster
19/01/31 18:43:19 INFO storage.BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
19/01/31 18:43:19 INFO storage.BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
19/01/31 18:43:19 INFO storage.DiskBlockManager: Created local directory at /PBMG/users/ggbmgphdpngrp/prime/cep/tmp/blockmgr-88cc1ce5-d255-4009-9864-25e5f567879e
19/01/31 18:43:19 INFO memory.MemoryStore: MemoryStore started with capacity 6.2 GB
19/01/31 18:43:20 INFO spark.SparkEnv: Registering OutputCommitCoordinator
19/01/31 18:43:20 INFO util.log: Logging initialized @2402ms
19/01/31 18:43:20 INFO server.Server: jetty-9.3.z-SNAPSHOT
19/01/31 18:43:20 INFO server.Server: Started @2475ms
19/01/31 18:43:20 INFO server.AbstractConnector: Started ServerConnector@3ad394e6{HTTP/1.1,[http/1.1]}{0.0.0.0:52000}
19/01/31 18:43:20 INFO util.Utils: Successfully started service 'SparkUI' on port 52000.
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@26f143ed{/jobs,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61a5b4ae{/jobs/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5b69fd74{/jobs/job,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@77b325b3{/jobs/job/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7e8e8651{/stages,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@271f18d3{/stages/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61e3a1fd{/stages/stage,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@315df4bb{/stages/stage/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5cad8b7d{/stages/pool,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@25243bc1{/stages/pool/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2e6ee0bc{/storage,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@467f77a5{/storage/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@420bc288{/storage/rdd,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@308a6984{/storage/rdd/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7a34b7b8{/environment,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3be8821f{/environment/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3b65e559{/executors,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@74a9c4b0{/executors/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1c05a54d{/executors/threadDump,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5fd9b663{/executors/threadDump/json,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@10567255{/static,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@60b85ba1{/,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@117632cf{/api,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@159e366{/jobs/job/kill,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@24528a25{/stages/stage/kill,null,AVAILABLE,@Spark}
19/01/31 18:43:20 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at http://10.6.209.22:52000
19/01/31 18:43:20 INFO spark.SparkContext: Added JAR file:/PBMG/users/ggbmgphdpngrp/prime/cep/prime-cep.jar at spark://10.6.209.22:50000/jars/prime-cep.jar with timestamp 1548956600301
19/01/31 18:43:20 INFO util.Utils: Using initial executors = 15, max of spark.dynamicAllocation.initialExecutors, spark.dynamicAllocation.minExecutors and spark.executor.instances
19/01/31 18:43:24 INFO yarn.Client: Requesting a new application from cluster with 8 NodeManagers
19/01/31 18:43:25 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (61440 MB per container)
19/01/31 18:43:25 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead
19/01/31 18:43:25 INFO yarn.Client: Setting up container launch context for our AM
19/01/31 18:43:25 INFO yarn.Client: Setting up the launch environment for our AM container
19/01/31 18:43:25 INFO yarn.Client: Preparing resources for our AM container
19/01/31 18:43:25 INFO security.HadoopFSCredentialProvider: getting token for: hdfs://nameservice-np/user/ggbmgphdpngrp
19/01/31 18:43:25 INFO hdfs.DFSClient: Created token for ggbmgphdpngrp: HDFS_DELEGATION_TOKEN owner=ggbmgphdpngrp@BMEDIA.BAGINT.COM, renewer=yarn, realUser=, issueDate=1548956605075, maxDate=1549561405075, sequenceNumber=1281621, masterKeyId=1013 on ha-hdfs:nameservice-np
19/01/31 18:43:26 INFO hive.metastore: Trying to connect to metastore with URI thrift://gtunxlnu00853.server.arvato-systems.de:9083
19/01/31 18:43:26 INFO hive.metastore: Opened a connection to metastore, current connections: 1
19/01/31 18:43:26 INFO hive.metastore: Connected to metastore.
19/01/31 18:43:27 INFO metadata.Hive: Registering function dateconversion com.infosys.bmg.analytics.Date_Convert
19/01/31 18:43:27 INFO metadata.Hive: Registering function calc_week com.bmg.main.CalcWeek
19/01/31 18:43:27 INFO metadata.Hive: Registering function prev_week com.infosys.bmg.analytics.HiveUdfPrevWeek
19/01/31 18:43:27 INFO metadata.Hive: Registering function prev_week com.infosys.bmg.analytics.HiveUdfPrevWeek
19/01/31 18:43:27 INFO metadata.Hive: Registering function dateconversion com.infosys.bmg.analytics.Date_Convert
19/01/31 18:43:27 INFO metadata.Hive: Registering function date_convert com.infosys.bmg.analytics.Date_Convert
19/01/31 18:43:27 INFO metadata.Hive: Registering function calc_week com.infosys.bmg.analytics.HiveUdfCalcWeek
19/01/31 18:43:27 INFO metadata.Hive: Registering function day_of_week com.infosys.bmg.analytics.HiveUdfDayOfWeek
19/01/31 18:43:27 INFO metadata.Hive: Registering function beginning_of_fin_week_func com.infosys.bmg.date.Begining_Of_Financial_Week
19/01/31 18:43:27 INFO metadata.Hive: Registering function end_of_fin_week_func com.infosys.bmg.date.End_Of_Financial_Week
19/01/31 18:43:27 INFO metadata.Hive: Registering function dateconversion com.infosys.bmg.analytics.DateConvertFlash
19/01/31 18:43:27 INFO metadata.Hive: Registering function beginning_of_fin_week_func_ada com.infosys.bmg.date.BeginingOfFinancialWeekADA
19/01/31 18:43:27 INFO metadata.Hive: Registering function end_of_fin_week_func_ada com.infosys.bmg.date.EndOfFinancialWeekADA
19/01/31 18:43:27 INFO metadata.Hive: Registering function first_financial_day_func_ada com.infosys.bmg.date.FirstFinancialDayOfYearADA
19/01/31 18:43:27 INFO metadata.Hive: Registering function titleconversionudf com.infosys.bmg.Pr