Member since
03-23-2021
10
Posts
1
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
803 | 07-23-2021 01:41 AM |
07-23-2021
01:41 AM
I fixed this error. 1. 'yarn' user Authenticate to kerberos as yarn/[HOST_NAME]@[REALM]. 2. YARN => CONFIGS => ADVANCED => Advanced container-executor => container-executor-configuration-template change the setting below banned.users=hdfs,mapred,bin
min.user.id={{min_user_id}} 3. YARN => CONFIGS => SETTINGS - Set 'Memory-Node' to fit my node. - Set 'Memory-Container-Minimum Container Size' to fiy my node. - Set 'Memory-Container-Maximum Container Size' to fiy my node.
... View more
07-20-2021
10:47 PM
hello. I'm using on apache ambari 2.7.5.0 and hdp 3.1.4.0.
After installing kerberos on ambari, hdfs, hive, and hbase were successfully authenticated using keytab and then used services.
However, if I use yarn to run wordcount, I get an error. In the case of the file permission executing wordcount, the same was confirmed, and in the case of the user name authenticated using keytab, the same was confirmed.
When running wordcount using yarn, the same error message as when yarn node -list -all is output is displayed. Below is the error message printed by 'yarn node -list -all.'
21/07/21 14:24:03 INFO retry.RetryInvocationHandler: java.io.IOException: DestHost:destPort hadoop1:8050 , LocalHost:localPort hadoop1/192.168.22.168:0. Failed on local exception: java.io.IOException: org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS], while invoking ApplicationClientProtocolPBClientImpl.getClusterNodes over null after 24 failover attempts. Trying to failover after sleeping for 17666ms.
21/07/21 14:24:21 WARN ipc.Client: Exception encountered while connecting to the server : org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS]
i'm using a single node for ambari.
Anyone know a solution to this problem?
... View more
Labels:
06-30-2021
10:37 PM
That was same my file. I changed '/var/lib/cloudera-scm-agent/uuid' file was not [noeol] file, so I changed that file. And can i show your error log in /var/run/cloudera-scm-agnet/process-*-HUE_SERVER Thank you
... View more
06-30-2021
06:05 PM
Hi @paliwalhadoop I changed /opt/cloudera/cm/bin/altscript.sh {{CMF_CONF_DIR}}/{{keystoreFileName}} to fit my environment. can you show your log or script file ? Thank you
... View more
04-05-2021
05:40 PM
Hey @vaishaakb thanks for your reply. I wasn't changed the CDH version. But i tested to change the path of the '/var/lib' folder location. Like a 'sevice-monitor', 'host-monitor'. And i current location recovery to '/var/lib'. In this host another roles started healthy. I check again stderr, stdout and i found a "altscript.sh" this file causing the problem. That problem is a not found a "creds.localjceks" file. So changed "/opt/~~~~/altscript.sh" the file. And i solved this problem. Thanks again for your reply.
... View more
03-30-2021
02:34 AM
I'm using CDH-6.2.0. and hue. HUE_LOAD_BALANCER is run fine. But HUE_SERVER does not run on each node(2 node). I've looked at the logs, but I can't find any problems. The strange thing is that even if you run the server once, it looks like it is running 4 times in the runcpserver.log as shown in the figure below, and in the supervisor log, you can see a log saying that it was executed too many times. That's my own log file. supervisord.log 2021-03-30 16:37:03,391 INFO spawned: '178-hue-HUE_SERVER' with pid 24378
2021-03-30 16:37:12,722 INFO exited: 178-hue-HUE_SERVER (exit status 0; not expected)
2021-03-30 16:37:13,726 INFO spawned: '178-hue-HUE_SERVER' with pid 24586
2021-03-30 16:37:22,823 INFO exited: 178-hue-HUE_SERVER (exit status 0; not expected)
2021-03-30 16:37:24,829 INFO spawned: '178-hue-HUE_SERVER' with pid 24857
2021-03-30 16:37:33,645 INFO exited: 178-hue-HUE_SERVER (exit status 0; not expected)
2021-03-30 16:37:36,653 INFO spawned: '178-hue-HUE_SERVER' with pid 25067
2021-03-30 16:37:45,512 INFO exited: 178-hue-HUE_SERVER (exit status 0; not expected)
2021-03-30 16:37:46,174 INFO gave up: 178-hue-HUE_SERVER entered FATAL state, too many start retries too quickly runcpserver.log [30/Mar/2021 16:37:11 ] settings INFO Welcome to Hue 4.3.0
[30/Mar/2021 16:37:21 ] settings INFO Welcome to Hue 4.3.0
[30/Mar/2021 16:37:32 ] settings INFO Welcome to Hue 4.3.0
[30/Mar/2021 16:37:44 ] settings INFO Welcome to Hue 4.3.0 When i started in Cloudera Manager. CM print that "Supervisor returned FATAL. Please check the role log file, stderr, or stdout." Thanks.
... View more
Labels:
- Labels:
-
Cloudera Hue
-
Cloudera Manager
03-23-2021
08:03 PM
1 Kudo
Thank you everyone. My problem is just uuid file in line feed.
... View more
03-23-2021
05:35 AM
Thank you. I'll check again CM server logs.
... View more
03-23-2021
05:33 AM
no. I've been using it well before, and I have this problem since I added something new to my topic.
... View more
03-23-2021
02:00 AM
hello i'm using CDH 6.2.0. I have a problem with the Cloudera management service not running. When I turn on the cloudera-scm-agent and cloudera-scm-server, it looks like the picture below. And when I run a service, I get a message saying "Role not started due to unhealthy host". this is my cloudera-scm-agent log. [23/Mar/2021 17:37:21 +0000] 3013 MonitorDaemon-Reporter throttling_logger ERROR Error sending messages to firehose: mgmt-HOSTMONITOR-6bd66bc3db6890d3e7e602f52c725199
Traceback (most recent call last):
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/cmf/monitor/firehose.py", line 121, in _send
self._port)
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/avro/ipc.py", line 469, in __init__
self.conn.connect()
File "/usr/lib64/python2.7/httplib.py", line 824, in connect
self.timeout, self.source_address)
File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
raise err
error: [Errno 111] Connection refused
[23/Mar/2021 17:46:27 +0000] 3013 MainThread heartbeat_tracker INFO HB stats (seconds): num:41 LIFE_MIN:0.02 min:0.01 mean:0.02 max:0.06 LIFE_MAX:0.13
[23/Mar/2021 17:47:21 +0000] 3013 MonitorDaemon-Reporter throttling_logger ERROR (9 skipped) Error sending messages to firehose: mgmt-HOSTMONITOR-6bd66bc3db6890d3e7e602f52c725199
Traceback (most recent call last):
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/cmf/monitor/firehose.py", line 121, in _send
self._port)
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/avro/ipc.py", line 469, in __init__
self.conn.connect()
File "/usr/lib64/python2.7/httplib.py", line 824, in connect
self.timeout, self.source_address)
File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
raise err
error: [Errno 111] Connection refused
[23/Mar/2021 17:56:27 +0000] 3013 MainThread heartbeat_tracker INFO HB stats (seconds): num:40 LIFE_MIN:0.01 min:0.01 mean:0.01 max:0.03 LIFE_MAX:0.13
[23/Mar/2021 17:57:21 +0000] 3013 MonitorDaemon-Reporter throttling_logger ERROR (9 skipped) Error sending messages to firehose: mgmt-HOSTMONITOR-6bd66bc3db6890d3e7e602f52c725199
Traceback (most recent call last):
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/cmf/monitor/firehose.py", line 121, in _send
self._port)
File "/opt/cloudera/cm-agent/lib/python2.7/site-packages/avro/ipc.py", line 469, in __init__
self.conn.connect()
File "/usr/lib64/python2.7/httplib.py", line 824, in connect
self.timeout, self.source_address)
File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
raise err
error: [Errno 111] Connection refused i don't know how to solve this problem. (add message) This problem occurs when the cloudera files in /var/lib are moved to another disk and linked. And Permisson is already changed.
... View more
Labels:
- Labels:
-
Apache Hadoop