Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 922 | 06-04-2025 11:36 PM | |
| 1523 | 03-23-2025 05:23 AM | |
| 752 | 03-17-2025 10:18 AM | |
| 2700 | 03-05-2025 01:34 PM | |
| 1800 | 03-03-2025 01:09 PM |
07-02-2018
08:09 PM
@M Sainadh Can you share the contents of /var/log/atlas/atlas.20180201-102756.err. Atlas need to access these 2 hbase tables atlas_titan, ATLAS_ENTITY_AUDIT_EVENTS, so you need to has hbase up and running . To check that the tables were created do the following $ cd /usr/hdp/current/zookeeper-client/bin/ Once connected to zookeeper look for the secure hbase as you have kerberos enabled [zk: localhost:2181(CONNECTED) 1] ls /hbase-secure/table
[ATLAS_ENTITY_AUDIT_EVENTS, hbase:meta, atlas_titan, hbase:namespace, hbase:acl] Above you see the tables if the 2 tables weren't created please use this How to manually create tables for Atlas in HBase to create the ATLAS_ENTITY_AUDIT_EVENTS and atlas_titan hbase tables HTH
... View more
07-02-2018
07:46 PM
@kanna k If the Active /standby namenodes switch status unexpectedly then have a look at the NTPD setup ! Are your node in sync? Kerberos is very sensitive with time so ensure your cluster time is in sync.
... View more
07-02-2018
11:38 AM
@kanna k You have a corrupt journalnode, Please follow this HCC doc to resolve the issue. Assuming that this is happening on a single JournalNode then you can try the following:
As a precaution, stop HDFS. This will shut down all Journalnodes as well. On the node in question, move the fsimage edits directory (/hadoop/hdfs/journal/xxxxx/current) to an alternate location. Copy the fsimage edits directory (/hadoop/hdfs/journal/xxxxx/current) from a functioning JournalNode to this node. Start HDFS. HTH
... View more
07-02-2018
07:24 AM
@n c Can you share you scheduler setting in YARN-->Configs-->Advanced-->Scheduler -->Capacity scheduler
... View more
07-01-2018
12:53 PM
6 Kudos
@harlem X What's a Keytab
Is a container to store Kerberos principal's keys. A key in terms of Kerberos can be some random bytes. You can use container to store a password for a user principal or a password (key) for a service principal.
Keytabs are cryptographic files containing an representation of the service and its long-term key.Keytabs are used to either de-crypt the Kerberos service ticket of a Service or user to an hadoop service or authenticate the service itself to another service on the network.
The permissions of the keytabs is usually 400 or 440 for services like (hdfs, hbase, ambari-qa, accumulo, and spnego). These keytabs are used by other services rather than just the owner for testing connectivity to the cluster, and other functions on startup. All of the other keytabs are only readable by the service account that owns the file and the group that owns the keytabs (hadoop) group, and should be reserved for service accounts.
What's a Kerberos service
Represents an application either acting as a client or as a server, it does not really matter which way. Such application may have own arrangement on how it runs (which UID/GID it uses on the operating system level) but it
is not important from Kerberos point of view because Kerberos is not responsible for the identity of your application (or user), it only deals with Kerberos principals and their keys. Headless keytab What you call 'headless keytab' is user principal keys. This already makes an assumption that you have a user principal that corresponds to certain POSIX user it will start services without prompting for password e.g(smokeuser,hdfs,hbase). Headless principals are not bound to a specific host or node, they have the syntax: - @ EXAMPLE.COM For Example: Headless: hdfs-mycluster@EXAMPLE.COM
Service keytab Service principal is something that does not need to be a POSIX user,they are mostly applications that have own arrangement on how they run on the OS level and need to interact with the Kerberized cluster. Kerberos services have traditional meaning e.g there could be host/fqdn REALM service principal that 'represents' this host in Kerberos realm. The same service principal may be shared by several applications: for example, both SSSD and SSH daemon use host/fqdn REALM for own needs. SSSD uses it as a client when authenticating against IPA LDAP server using SASL GSSAPI, and SSH daemon uses host/fqdn REALM key to represent itself as a server to incoming SSH
clients using GSSAPI or Kerberos authentication methods. Service principals are bound to a specific service and host or node, they have the syntax: / @ EXAMPLE.COM e.g: Service: nn/c6601.ambari.apache.org@EXAMPLE.COM For example in FreeIPA it is recommended to create service principals to represent applications as they are not required to have global POSIX identity associated with them and they are usually running on a specific host. In addition, if they are accepting SASL GSSAPI authentication method to access themselves, a client application will usually build up a target principal based on the hostname they run on, e.g. HTTP/fqdn REALM for a web server running on the host fqdn. Thus, there is clear arrangement between client and server applications on what they expect from each other on Kerberos (or SASL GSSAPI) level. User principals For user principals you can store user's password in a 'headless keytab' to allow some impersonation of the user for certain needs but it is irrelevant from Kerberos level what identity is there, both service and user principals can equally be used at a client side to initiate authentication towards a server. Script for generating headless keytabs Generating a headless keytab uses the same utility using kadmin.local you just specify headless assuming you have already a POSIX see above. kadmin: ktadd -k /etc/security/keytabs/solr-headless-keytab host/prodsolr.exmple.com The above will generate solr-headless-keytab for host prodsolr HTH
... View more
06-29-2018
09:33 PM
1 Kudo
@Steve Kiaie A simple answer is NO you cannot remove the SNN in a non NameNode HA setup reasons below: Secondary NameNode(SNN) and Standby NameNode are mutually exclusive,enabling HA is not mandatory. But, when it is enabled, you can't use Secondary Namenode. So, either Secondary Namenode is enabled OR Standby Namenode is enabled. Secondary NameNode downloads the FsImage and EditLogs from the NameNode and then it merges EditLogs with the Fsimage periodically. It keeps edits log size within a limit. After that, it stores the modified FsImage into persistent storage. So we can use FsImage in case of NameNode failure. Secondary namenode is just a helper for Namenode. It gets the edit logs from the namenode in regular intervals and applies to fsimage.
Once it has new fsimage, it copies back to namenode. Namenode will use this fsimage for the next restart, which will reduce the startup time. Secondary Namenode's whole purpose is to have a checkpoint in HDFS. Its just a helper node for namenode. That’s why it also known also as checkpoint node,But, It can't replace namenode on namenode's failure.NameNode is single point of Failure (SPOF). If namenode fails, all clients would unable to read/write files. In such event, whole cluster will be out of service until new namenode is up. The standby namenode provides automatic failover in case Active Namenode (can be simply called 'Namenode' if HA is not enabled) fails.You need a Zookeeper cluster (quorum of 3) to add a Standby namenode hence enabling NameNode HA
To overcome this issue; Standby NameNode comes into picture.
It does three things: Merging fsimage and edits-log files. (Secondary-namenode's work). Receive online updates of the file system meta-data using journalnoodes. Apply the changes to its memory state and persist them on disks just like the name-node does. Thus at any time the Backup node contains an up-to-date image of the namespace both in memory and on local disk(s).
Cluster will switch over to the new name-node (this standby-node) if the active namenode dies
High availability feature provides an extra NameNode to hadoop architecture,this feature provides automatic failover. If active NameNode fails, then standby-Namenode takes all the responsibility of active node. And cluster continues to work. I have attached 2 visuals to help you understand HTH
... View more
06-28-2018
03:10 PM
@kiwi z Could you share your solution so that other members who encounter the same situation could have a quick solution
... View more
06-27-2018
10:06 PM
@kiwi z Can you see any error messages in /var/log/ranger/kms/catalina.out thats the startup logfile. If you see message about InvalidKeyException like below java.security.InvalidKeyException:Illegal key size
at javax.crypto.Cipher.checkCryptoPerm(Cipher.java:1039) That indicates JCE issue so install JDK JCE and that should resolve the issue distribute and install JCE on all nodes. HTH
... View more
06-27-2018
05:26 PM
@Samant Thakur Good to know,give yourself the points then close this thread.
... View more
06-27-2018
11:43 AM
@Michael Bronson Can you share your postgresql.conf ? Remember to protect vital info
... View more