Member since
04-11-2016
535
Posts
148
Kudos Received
77
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7497 | 09-17-2018 06:33 AM | |
1827 | 08-29-2018 07:48 AM | |
2744 | 08-28-2018 12:38 PM | |
2126 | 08-03-2018 05:42 AM | |
1991 | 07-27-2018 04:00 PM |
07-26-2017
06:08 AM
1 Kudo
@Dhiraj Then, you need to run 'Drop database cascade'.
... View more
07-07-2017
05:54 AM
If the mysqldump is for different version other than Hive 2.1.1000 in HDP 2.6. Then do the following: 1. Stop Hive services from Ambari. 2. Create new database under MySQL as say hive2: mysql> create database hive2;
Query OK,1 row affected (0.00 sec)
mysql> grant all privileges on hive2.* to 'hive'@'%' identified by'hive';
Query OK,0 rows affected (0.00 sec) 3. Restore database as: mysql -u hive -phive hive2 < dumpfilename.sql 4. Update database connection string for mysql under Ambari -> Hive configs. 5. Save configurations and try restarting. Since there is different in VERSION, service startup would fail. 6. Run Hive schematool command to upgrade the schema as below: hive@ssnode260 bin]$ /usr/hdp/2.6.0.3-8/hive2/bin/schematool -upgradeSchema -dbType mysql 7. Restart Hive services from Ambari. If the Hive metadata version is same as Hive 2.1.1000 in HDP 2.6, then follow steps 1 through 5.
... View more
Labels:
07-07-2017
08:54 AM
Thank you for your answer but I think I was not clear enough with my explanation. The MySQL dump come from MariaDB. And I want to ingest the data from MariaDB but I can't access directly MariaDB but only a mysqldump of this database. I want to recreate the MariaDB in HDP 2.6 in some way. My question is the same what are the best scenarios to do this workflow ? Thank you a lot again. Arnault
... View more
07-05-2017
08:33 AM
@Anurag Mishra The fsimage would be under <namenode_dir>/current.
... View more
06-01-2017
09:05 PM
1 Kudo
And the solution appears to be we had some files .../hadoop/yarn/local/usercache/hive/filecache with odd permissions. Flushing that directory on the data nodes and the query runs.
... View more
05-31-2017
07:28 AM
@mÁRIO Rodrigues
This seems to be known issue and Jira DRILL-5214 has been raised for the same.
... View more
02-21-2017
07:59 AM
SYMPTOM When running Hive queries from Resource Manager log the following error is displayed: 2017-02-07 15:08:32,140 ERROR impl.MetricsSinkAdapter (MetricsSinkAdapter.java:publishMetricsFromQueue(148)) - Got sink exception, retry in 4600ms
org.apache.hadoop.metrics2.MetricsException: Failed to putMetrics
at org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink.putMetrics(HadoopTimelineMetricsSink.java:216)
at org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:186)
at org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:43)
at org.apache.hadoop.metrics2.impl.SinkQueue.consumeAll(SinkQueue.java:87)
at org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.publishMetricsFromQueue(MetricsSinkAdapter.java:134)
at org.apache.hadoop.metrics2.impl.MetricsSinkAdapter$1.run(MetricsSinkAdapter.java:88)
Caused by: java.net.UnknownHostException: http
at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:178)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:579)
at java.net.Socket.connect(Socket.java:528)
at java.net.Socket.<init>(Socket.java:425)
at java.net.Socket.<init>(Socket.java:280) ROOT CAUSE This issue occurs when reverse lookup returns incorrect hostname for the IP address. RESOLUTION To resolve this issue, fix the DNS issue under /etc/resolv.conf.
... View more
Labels:
06-12-2019
11:31 AM
i need some help with this error??!!!! hive> (base) hadoop@sajanraj:~$ schematool -dbType mysql -initSchema --verbose SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/usr/local/apache-hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] Metastore connection URL: jdbc:mysql://192.168.5.249:3306/hive?createDatabaseIfNotExist=true Metastore Connection Driver : com.mysql.jdbc.Driver Metastore connection User: hiveuser Wed Jun 12 14:34:03 IST 2019 WARN: Establishing SSL connection without server's identity verification is not recommended. According to MySQL 5.5.45+, 5.6.26+ and 5.7.6+ requirements SSL connection must be established by default if explicit option isn't set. For compliance with existing applications not using SSL the verifyServerCertificate property is set to 'false'. You need either to explicitly disable SSL by setting useSSL=false, or set useSSL=true and provide truststore for server certificate verification. Starting metastore schema initialization to 3.1.0 Initialization script hive-schema-3.1.0.mysql.sql Connecting to jdbc:mysql://192.168.5.249:3306/hive?createDatabaseIfNotExist=true Wed Jun 12 14:34:03 IST 2019 WARN: Establishing SSL connection without server's identity verification is not recommended. According to MySQL 5.5.45+, 5.6.26+ and 5.7.6+ requirements SSL connection must be established by default if explicit option isn't set. For compliance with existing applications not using SSL the verifyServerCertificate property is set to 'false'. You need either to explicitly disable SSL by setting useSSL=false, or set useSSL=true and provide truststore for server certificate verification. Connected to: MySQL (version 5.7.26-0ubuntu0.18.04.1) Driver: MySQL Connector Java (version mysql-connector-java-5.1.45 ( Revision: ${revinfo.commit} )) Transaction isolation: TRANSACTION_READ_COMMITTED 0: jdbc:mysql://192.168.5.249:3306/hive> !autocommit on Autocommit status: true 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @OLD_CHARACTER_SET_CLIENT=@@CHARACTER_SET_CLIENT */ No rows affected (0.002 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @OLD_CHARACTER_SET_RESULTS=@@CHARACTER_SET_RESULTS */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @OLD_COLLATION_CONNECTION=@@COLLATION_CONNECTION */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET NAMES utf8 */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40103 SET @OLD_TIME_ZONE=@@TIME_ZONE */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40103 SET TIME_ZONE='+00:00' */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40014 SET @OLD_UNIQUE_CHECKS=@@UNIQUE_CHECKS, UNIQUE_CHECKS=0 */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40014 SET @OLD_FOREIGN_KEY_CHECKS=@@FOREIGN_KEY_CHECKS, FOREIGN_KEY_CHECKS=0 */ No rows affected (0.005 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @OLD_SQL_MODE=@@SQL_MODE, SQL_MODE='NO_AUTO_VALUE_ON_ZERO' */ No rows affected (0.002 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40111 SET @OLD_SQL_NOTES=@@SQL_NOTES, SQL_NOTES=0 */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @saved_cs_client = @@character_set_client */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = utf8 */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> CREATE TABLE IF NOT EXISTS `BUCKETING_COLS` ( `SD_ID` bigint(20) NOT NULL, `BUCKET_COL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, `INTEGER_IDX` int(11) NOT NULL, PRIMARY KEY (`SD_ID`,`INTEGER_IDX`), KEY `BUCKETING_COLS_N49` (`SD_ID`), CONSTRAINT `BUCKETING_COLS_FK1` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`) ) ENGINE=InnoDB DEFAULT CHARSET=latin1 No rows affected (0.007 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = @saved_cs_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @saved_cs_client = @@character_set_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = utf8 */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> CREATE TABLE IF NOT EXISTS `CDS` ( `CD_ID` bigint(20) NOT NULL, PRIMARY KEY (`CD_ID`) ) ENGINE=InnoDB DEFAULT CHARSET=latin1 No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = @saved_cs_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @saved_cs_client = @@character_set_client */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = utf8 */ No rows affected (0 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> CREATE TABLE IF NOT EXISTS `COLUMNS_V2` ( `CD_ID` bigint(20) NOT NULL, `COMMENT` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, `COLUMN_NAME` varchar(767) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, `TYPE_NAME` MEDIUMTEXT DEFAULT NULL, `INTEGER_IDX` int(11) NOT NULL, PRIMARY KEY (`CD_ID`,`COLUMN_NAME`), KEY `COLUMNS_V2_N49` (`CD_ID`), CONSTRAINT `COLUMNS_V2_FK1` FOREIGN KEY (`CD_ID`) REFERENCES `CDS` (`CD_ID`) ) ENGINE=InnoDB DEFAULT CHARSET=latin1 No rows affected (0.002 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = @saved_cs_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET @saved_cs_client = @@character_set_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = utf8 */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> CREATE TABLE IF NOT EXISTS `DATABASE_PARAMS` ( `DB_ID` bigint(20) NOT NULL, `PARAM_KEY` varchar(180) CHARACTER SET latin1 COLLATE latin1_bin NOT NULL, `PARAM_VALUE` varchar(4000) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT NULL, PRIMARY KEY (`DB_ID`,`PARAM_KEY`), KEY `DATABASE_PARAMS_N49` (`DB_ID`), CONSTRAINT `DATABASE_PARAMS_FK1` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`) ) ENGINE=InnoDB DEFAULT CHARSET=latin1 No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> /*!40101 SET character_set_client = @saved_cs_client */ No rows affected (0.001 seconds) 0: jdbc:mysql://192.168.5.249:3306/hive> CREATE TABLE `CTLGS` ( `CTLG_ID` BIGINT PRIMARY KEY, `NAME` VARCHAR(256), `DESC` VARCHAR(4000), `LOCATION_URI` VARCHAR(4000) NOT NULL, UNIQUE KEY `UNIQUE_CATALOG` (`NAME`) ) ENGINE=InnoDB DEFAULT CHARSET=latin1 Error: Table 'CTLGS' already exists (state=42S01,code=1050) Closing: 0: jdbc:mysql://192.168.5.249:3306/hive?createDatabaseIfNotExist=true org.apache.hadoop.hive.metastore.HiveMetaException: Schema initialization FAILED! Metastore state would be inconsistent !! Underlying cause: java.io.IOException : Schema script failed, errorcode 2 org.apache.hadoop.hive.metastore.HiveMetaException: Schema initialization FAILED! Metastore state would be inconsistent !! at org.apache.hive.beeline.HiveSchemaTool.doInit(HiveSchemaTool.java:594) at org.apache.hive.beeline.HiveSchemaTool.doInit(HiveSchemaTool.java:567) at org.apache.hive.beeline.HiveSchemaTool.main(HiveSchemaTool.java:1517) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:323) at org.apache.hadoop.util.RunJar.main(RunJar.java:236) Caused by: java.io.IOException: Schema script failed, errorcode 2 at org.apache.hive.beeline.HiveSchemaTool.runBeeLine(HiveSchemaTool.java:1226) at org.apache.hive.beeline.HiveSchemaTool.runBeeLine(HiveSchemaTool.java:1204) at org.apache.hive.beeline.HiveSchemaTool.doInit(HiveSchemaTool.java:590) ... 8 more *** schemaTool failed ***
... View more
11-07-2017
06:22 PM
I would like to know the same - if there is way to change the default port via the ambari bootstrap instead of the UI wizard.
... View more
08-23-2017
01:30 PM
Issue is related to hadoop.proxyuser.<user_name>.hosts and hadoop.proxyuser.<user_name>.groups configurations from core-site under HDFS configs, for all properties hadoop.proxyuser.<user_name>.hosts changed values to * and it will worked.
... View more