Member since
05-11-2018
10
Posts
0
Kudos Received
0
Solutions
06-21-2018
03:06 AM
Basically issue is the Hadoop is running an id command against the user and not including -- in front of the user if the username starts with a -. 1. As per POSIX standard, a username should not start with a hyphen character, and "-". It is causing this problem. 2. The authorization is done over Kerberos which is working fine. 3. Seems like this userid needs its Unix groups to perform some action on HDFS which is throwing this error (eg writing to a directory using group permissions). This is because the group lookup is not working due to above mentioned issue.
... View more
05-17-2018
11:09 AM
If I am using -- option from the command line, I am getting the group details. I am not sure why this error is being logged. Any thoughts? # groups -- -SERV-CHEF groups: -SERV-CHEF: no such user # groups -- -SERV-CHEF@domain success!
... View more
05-17-2018
07:15 AM
We are getting continuous PartialGroupNameException for a specific user account. Below is the sample error snippet. ++++++ WARN org.apache.hadoop.security.ShellBasedUnixGroupsMapping: unable to return groups for user -SERV-CHEF PartialGroupNameException The user name '-SERV-CHEF' is not found. id: invalid option -- 'S' Try 'id --help' for more information. Try 'id --help' for more information. id: invalid option -- 'S' Try 'id --help' for more information. at org.apache.hadoop.security.ShellBasedUnixGroupsMapping.resolvePartialGroupNames(ShellBasedUnixGroupsMapping.java:212) at org.apache.hadoop.security.ShellBasedUnixGroupsMapping.getUnixGroups(ShellBasedUnixGroupsMapping.java:133) at org.apache.hadoop.security.ShellBasedUnixGroupsMapping.getGroups(ShellBasedUnixGroupsMapping.java:72) at org.apache.hadoop.security.Groups$GroupCacheLoader.fetchGroupList(Groups.java:368) at org.apache.hadoop.security.Groups$GroupCacheLoader.load(Groups.java:309) at org.apache.hadoop.security.Groups$GroupCacheLoader.load(Groups.java:267) at com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3568) at com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2350) at com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2313) at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228) at com.google.common.cache.LocalCache.get(LocalCache.java:3965) at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3969) at com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4829) at org.apache.hadoop.security.Groups.getGroups(Groups.java:225) at org.apache.hadoop.security.UserGroupInformation.getGroups(UserGroupInformation.java:1778) at org.apache.hadoop.security.UserGroupInformation.getGroupNames(UserGroupInformation.java:1766) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.(FSPermissionChecker.java:66) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.getPermissionChecker(FSDirectory.java:3468) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getPermissionChecker(FSNamesystem.java:4079) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4269) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:901) at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getFileInfo(AuthorizationProviderProxyClientProtocol.java:528) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:839) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2216) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2212) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2210) ++++++ Source : /logs/hadoop-hdfs/hadoop-cmf-CD-HDFS-pFXTJXEX-NAMENODE-.log.out Command output from name node: # id -Gn -SERV-CHEF id: invalid option -- 'S' Try 'id --help' for more information. You have mail in /var/spool/mail/root # # hdfs groups -SERV-CHEF -SERV-CHEF : # Could you help me here with some insights, please?
... View more
Labels:
- Labels:
-
HDFS
05-15-2018
06:32 AM
I am using Splunk in my environment. Is it being logged anywhere while a query is spilling to disk? If yes, I can create an alert from Splunk to notify me the query details while spilling to disk happens.
... View more
05-15-2018
04:58 AM
Hi Tim, Thank you for your inputs. I have checked the matrics links given. Is unit "queries per second" is a list of individual queries or the count of the queries? ex: Metric Name Description Unit Parents CDH Version queries_spilled_memory_rate Impala queries that spilled to disk queries per second cluster CDH 5 I need to track specific queries that are spilling to disk. Means if I am enabling the spill-to-disk option, I need to get an alert if a specific query is spilling memory to disk with the query details so that I can notify the owner of that query. I can set the scratch limit to a specific value to control spill space usage. However, need to track each and every query that spills to disk.
... View more
05-14-2018
08:12 AM
Is spill-to-disk being logged in any logs? if yes, I can set up an alert from the Splunk.
... View more
05-14-2018
04:37 AM
Thanks Tim for your reply. The first option will display the details of the memory spilled per query from the impala query section if that attribute is selected to be displayed. The second option under the utilization section will give us the details of average spill and maximum spill per resource pool. My requirement is that, if I am able enabling spill to disk feature in my cluster, I want to be notified if any spill to disk is happening. Do we have any option in cloudera manager to create this alert?
... View more
05-11-2018
07:28 AM
Hello Team, As per Impala release notes for Impala 2.5, +++++ Spill-to-disk feature now always recommended. In earlier releases, the spill-to-disk feature could be turned off using a pair of configuration settings, enable_partitioned_aggregation=false and enable_partitioned_hash_join=false. The latest improvements in the spill-to-disk mechanism, and related features that interact with it, make this feature robust enough that disabling it is now no longer needed or supported. In particular, some new features in Impala 2.5 and higher do not work when the spill-to-disk feature is disabled. +++++ If spill-to-disk is enabled, is there an option to monitor the spill-to-disk instances so that I can monitor the query that is causing it.
... View more
Labels:
- Labels:
-
Apache Impala
-
Cloudera Manager