Member since
01-19-2017
3676
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 557 | 06-04-2025 11:36 PM | |
| 1111 | 03-23-2025 05:23 AM | |
| 561 | 03-17-2025 10:18 AM | |
| 2110 | 03-05-2025 01:34 PM | |
| 1319 | 03-03-2025 01:09 PM |
09-13-2019
05:51 AM
@kvinod Good then close the thread. It's always a good idea to have an external database to host not only the CM/Ambari database but also Oozie, hive, ranger etc in a multitenant architecture and maybe backup the database host periodically to avoid having your backend databases all over the cluster. 20 cents advice !
... View more
09-13-2019
05:45 AM
@kvinod It seems to me you first changed the memory parameters and when you switch user the environmental settings are lost. Could you try these steps below $ sudo -u hbase $ export _JAVA_OPTIONS="-Xmx2048m -Xms2048m" $ hbase hbck Please adjust the java options according to the memory available on your cluster.
... View more
09-12-2019
11:53 AM
@kvinod That error "java.lang.OutOfMemoryError: Java heap space " is as it says memory-related hbase backup [hbck] will acquire by default the heap size of JVM which is usually ~1GB. My config is single cluster 2.6.5 HDP with 12 GB memory You can try setting below option and then try executing hbck adjust it according to the memory available. export _JAVA_OPTIONS="-Xmx2048m -Xms2048m" Make sure your run the backup as user hbase. [hbase@nanyuki ~]$ hbase hbck See attached output [hbase@nanyuki ~]$ hbase hbck 2019-09-12 20:33:35,729 INFO [main] zookeeper.RecoverableZooKeeper: Process identifier=hbase Fsck connecting to ZooKeeper ensemble=nanyuki.socar.de:2181 2019-09-12 20:33:35,770 INFO [main] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-292--1, built on 05/11/2018 07:15 GMT 2019-09-12 20:33:35,770 INFO [main] zookeeper.ZooKeeper: Client environment:host.name=nanyuki.socar.de 2019-09-12 20:33:35,770 INFO [main] zookeeper.ZooKeeper: Client environment:java.version=1.8.0_112 2019-09-12 20:33:35,770 INFO [main] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation 2019-09-12 20:33:35,770 INFO [main] zookeeper.ZooKeeper: Client environment:java.home=/usr/jdk64/jdk1.8.0_112/jre 2019-09-12 20:33:35,771 INFO [main] zookeeper.ZooKeeper: Client environment:java.class.path=/usr/hdp/2.6.5.0-292/hbase/conf:/usr/jdk64/jdk1.8.0_112/lib/tools.jar:/usr/hdp/2.6.5.0-292/hbase:/usr/hdp/2.6.5.0-292/hbase/lib/activation-1.1.jar:/usr/hdp/2.6.5.0-292/hbase/lib/aopalliance-1.0.jar:/usr/hdp/2.6.5.0-292/hbase/lib/apacheds-i18n-2.0.0-M15.jar .... ... 2019-09-12 20:33:41,203 INFO [main] util.HBaseFsck: Computing mapping of all store files 2019-09-12 20:33:41,398 INFO [main] util.HBaseFsck: Validating mapping using HDFS state 2019-09-12 20:33:41,398 INFO [main] util.HBaseFsck: Computing mapping of all link files 2019-09-12 20:33:41,503 INFO [main] util.HBaseFsck: Validating mapping using HDFS state 2019-09-12 20:33:42,148 INFO [main] util.HBaseFsck: Finishing hbck Summary: Table testtable3 is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table hbase:meta is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table hbase:acl is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table ATLAS_ENTITY_AUDIT_EVENTS is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table hbase:backup is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table PDFTable is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table jina is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table hbase:namespace is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table PDFTable:DOCID001 is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 Table atlas_titan is okay. Number of regions: 1 Deployed on: nanyuki.socar.de,16020,1568312911672 0 inconsistencies detected. Status: OK 2019-09-12 20:33:42,153 INFO [main] zookeeper.ZooKeeper: Session: 0x16d26b3cb56003c closed 2019-09-12 20:33:42,153 INFO [main] client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2019-09-12 20:33:42,153 INFO [main] client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x16d26b3cb56003d 2019-09-12 20:33:42,155 INFO [main-EventThread] zookeeper.ClientCnxn: EventThread shut down 2019-09-12 20:33:42,157 INFO [main] zookeeper.ZooKeeper: Session: 0x16d26b3cb56003d closed 2019-09-12 20:33:42,157 INFO [main-EventThread] zookeeper.ClientCnxn: EventThread shut down [hbase@nanyuki ~]$ So the backup run smoothly
... View more
09-11-2019
02:05 PM
1 Kudo
@Manoj690 Those are warning you can ignore! If indeed your HBase indeed is not running, please can you validate your steps against the install steps documented below. Here are the steps to how install standalone hbase
... View more
09-11-2019
12:17 PM
@ranger Can you set the ATLAS_HOME_DIR in atlas-env.sh, and other settings below by editing the hive-site.xml adding the below changes Set-up atlas hook in hive-site.xml of your hive configuration: <property> <name>hive.exec.post.hooks</name> <value>org.apache.atlas.hive.hook.HiveHook</value> </property> And <property> <name>atlas.cluster.name</name> <value>primary</value> </property> Add 'export HIVE_AUX_JARS_PATH=<atlas package>/hook/hive' in hive-env.sh of your hive configuration Copy <atlas-conf>/atlas-application.properties to the hive conf directory. cp /usr/hdp/<VERSION>/atlas/conf/atlas-application.properties to /usr/hdp/<VERSION>/hive/conf Then run the import locate /usr/hdp/<VERSION>/atlas/hook-bin/import-hive.sh Please let me know the outcome
... View more
09-10-2019
02:31 PM
1 Kudo
@ranger There are three modes for Hive Metastore deployment: Embedded Metastore Not recommended for production. Local Metastore This mode allows us to have many Hive sessions i.e. many users can use the metastore at the same time. It's achieved by using any JDBC compliant like MySQL. In this case, the javax.jdo.option.ConnectionURL property is set to jdbc:mysql://host/dbname? createDatabaseIfNotExist=true, and javax.jdo.option.ConnectionDriverName is set to com.mysql.jdbc.Driver. The JDBC driver JAR file for MySQL must be on Hive’s classpath, Remote Metastore In this mode, metastore runs on its own separate JVM, not in the Hive service JVM. If other processes want to communicate with the metastore server they can communicate using Thrift Network APIs here you have the ability to have one more metastore servers in this case to provide High availability. having said that it seems you are trying to use embedded Metastore. What I advice you to do is create one as the root through the Ambari ui, it will ask you for the DBName and Host which would be where you installed the MySQL database else pre-create the metastore. The Hive database must be created before loading the Hive database schema that explains why you are getting the startup error. Using Hive with MySQL On the Ambari Server host, stage the appropriate MySQL connector for later deployment. On the Ambari Server host, Download the MySQL Connector/JDBC driver from MySQL. Runambari-server setup --jdbc-db=mysql --jdbc-driver=/path/to/mysql/mysql-connector-java.jar Confirm that mysql-connector-java.jar is in the Java share directory. ls /usr/share/java/mysql-connector-java.jar Make sure the .jar file has the appropriate permissions - 644. Execute the following command: ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar Create a user for Hive and grant it permissions. using the MySQL database admin utility: # mysql -u root -p
CREATE USER '[HIVE_USER]'@'localhost' IDENTIFIED BY '[HIVE_PASSWORD]';
GRANT ALL PRIVILEGES ON *.* TO '[HIVE_USER]'@'localhost';
CREATE USER '[HIVE_USER]'@'%' IDENTIFIED BY '[HIVE_PASSWORD]';
GRANT ALL PRIVILEGES ON *.* TO '[HIVE_USER]'@'%';
CREATE USER '[HIVE_USER]'@'[HIVE_METASTORE_FQDN]' IDENTIFIED BY '[HIVE_PASSWORD]';
GRANT ALL PRIVILEGES ON *.* TO '[HIVE_USER]'@'[HIVE_METASTORE_FQDN]';
FLUSH PRIVILEGES; Where [HIVE_USER] is your desired Hive user name, [HIVE_PASSWORD] is your desired Hive user password and [HIVE_METASTORE_FQDN] is the Fully Qualified Domain Name of the Hive Metastore host. Create the Hive database. The Hive database must be created before loading the Hive database schema. # mysql -u root -p
CREATE DATABASE [HIVE_DATABASE] Where [HIVE_DATABASE] is your desired Hive database name. After the above step then in the Ambari UI when you reach the Hive Metastpre configuring stage use the same credentials and the "test" should succeed and that should fire up when you start all the HDP components HOpe that helps
... View more
09-08-2019
05:08 AM
1 Kudo
@shashank_naresh The compatibility states that KNIME Big Data Connectors are certified by Cloudera for CDH 5.x, by Hortonworks for HDP 2.1 to 2.4 but should be valid also for later versions as well as MapR for MapR 4.1. with Hive 0.13. It seems you are trying to access Knime UI withing HDP, that is not possible, the essence is to connect and work with the data. So the most appropriate access required is to the DWH over HDFS which in your case is Hive. Question How are you trying to connect to Knime? You should use JDBC check Big Data Extensions can be purchased at http://www.knime.org/knime-big-data-extensions. Refs: https://www.knime.com/knime-big-data-connectors https://hortonworks.com/wp-content/uploads/2014/12/Knime-Hortonworks-Solutions-Brief.pdf Hope that helps
... View more
09-01-2019
10:05 AM
@kvinod You can use MySQL for the component please follow https://www.cloudera.com/documentation/enterprise/latest/topics/cm_ig_mysql.html
... View more
08-31-2019
11:44 PM
@EricL Sorry about the confusion, the location of /etc/security/keytabs/* is specific to HDP the equivalent of /var/run/cloudera-scm-agent/process so try to map so the command $ klist -kt /var/run/cloudera-scm-agent/process/*.keytab shoule be valid
... View more
08-31-2019
12:35 PM
@kal I have written an article in response to a similar question in HCC before, you have a couple of things to do before this can work! My 2 cents advice you should have first configured the 3 clusters using single KDC, then added a second KDC's on each cluster then proceed to configure Cross_Realm Trust. In the process, you would have gained some knowledge on the implementation having said that you will need KDCs in a Master/Slave configuration to be able to propagate principal and Keytabs between the 2 KDC using krb5_prop this is another chapter on its own. I have already noticed also an error in your krb5.conf in the below part, first you have 3 KDC's which are not replicating.there are specific steps to enable replication between KDC's see the one highlighted in ORANGE the values on the left in [lower case ] should mirror on the right in [upper case] see valid example in BLUE [domain_realm] .xxxxxxx.net = XXXXXXX.NET xxxxxxx.net = XXXXXXX.NET .xxxxxxx.com = XXXXXXX.NET xxxxxxx.com = XXXXXXX.NET .sys.xxxxxxx.net = SPECTRA.XXXXXXX.NET sys.xxxxxx.net = SPECTRA.xxxxxx.NET --Valid-- [domain_realm] .xxxxxxx.net = XXXXXXX.NET xxxxxxx.net = XXXXXXX.NET .xxxxxxx.com = XXXXXXX.COM xxxxxxx.com = XXXXXXX.COM .spectra.xxxxxxx.net = SPECTRA.XXXXXXX.NET spectra.xxxxxx.net = SPECTRA.xxxxxx.NET You should also configure /etc/hosts file on all the 3 clusters to have IP--HOSTNAME--ALIAS and the files should be copied to all hosts in the cluster if DNS is not resolving. The example below depicts hosts in the 3 different network segments # Cluster 1 192.168.0.1 node1.SPECTRA.XXXXXXX.NET node1 192.168.0.2 node2.SPECTRA.XXXXXXX.NET node2 .... 192.168.0.3 node3.SPECTRA.XXXXXXX.NET node3 # Cluster 2 192.168.1.10 node01.XXXXXXX.NET node01 192.168.1.20 node02.XXXXXXX.NET node02 ......... 192.168.1.30 node03.XXXXXXX.NET node01 # Cluster 3 192.168.2.30 nodex.XXXXXXX.COM nodex 192.168.2.40 nodey.XXXXXXX.COM nodey ......... 192.168.2.50 nodez.XXXXXXX.COM nodez if you could breakdown your steps it would be easier to achieve as I reiterated in the beginning but its doable task. Please let me know
... View more