Member since
09-25-2015
356
Posts
382
Kudos Received
62
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1259 | 11-03-2017 09:16 PM | |
1059 | 10-17-2017 09:48 PM | |
1693 | 09-18-2017 08:33 PM | |
1920 | 08-04-2017 04:14 PM | |
2034 | 05-19-2017 06:53 AM |
08-04-2017
04:14 PM
1 Kudo
The above error is more indicative of a corrupted instal of the hive client. You mentioned that you are able to connect with LDAP userid and password. In the case where you are getting the error did you launch the beeline cli from the same machine where you launched when you were successfully able to connect using LDAP credentials?
... View more
08-03-2017
07:37 PM
1 Kudo
Can you also post application logs for application_1501335737720_0013?
... View more
08-01-2017
08:22 PM
1 Kudo
Can you post the Ambari console log for the operation and application logs for application_1501335737720_0014?
... View more
08-01-2017
08:18 PM
1 Kudo
What is your Sqoop command looks like? Are you using a custom DB query to pull the data and is that syntactically correct?
... View more
07-27-2017
11:03 PM
1 Kudo
Can you check and post what HIVE_AUX_JARS_PATH is set to in /etc/hive/conf/hive-env.sh?
... View more
07-24-2017
10:47 PM
2 Kudos
Running 2 Metastore/HS2 with different config is possible. If your cluster is Ambari managed, you could potentially achieve that by using Ambari Config Groups.
... View more
07-24-2017
09:58 PM
2 Kudos
Are you interacting with an ACID table in non-ACID mode? This can happen, you could try restarting the Hive services after enabling ACID (setting hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager) and then try the query.
... View more
07-17-2017
06:22 PM
1 Kudo
As you mentioned that you want 4 instances of HS2, 2 in binary HA and 2 in http HA, you can can acheive that by registering similar transport protocols with the same zookeeper namespace. So, in case of the two binary HA instances, you can define the configs like this hive.server2.zookeeper.namespace=hiveserver2-binary and for the other two http HA instances hive.server2.zookeeper.namespace=hiveserver2-http Then in the connection URL JDBC clients would provide the appropriate zkNamespace jdbc:hive2://$ZK_URL/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2-<protocol>
... View more
07-07-2017
11:18 PM
1 Kudo
Typically beeline -u "<jdbc_url>" -n <username> -w <password>, note the jdbc_url is quoted. Are you sure you did that?
... View more
07-05-2017
07:36 PM
1 Kudo
You need to catch the appropriate exception as the error indicates. Here is an example: try {
Class.forName(driverName);
} catch (ClassNotFoundException ex) {
ex.printStackTrace();
}
... View more
06-30-2017
05:19 AM
1 Kudo
Other than what @Ishan mentioned, multiple HiveServer2 and Metastore are recommended generally in a High Availability setup. This way when one of the services goes down the other service is available for the clients to connect to. Typically edge nodes are client facing nodes, so the client configs and tools are installed on them rather than services.
... View more
06-30-2017
04:52 AM
1 Kudo
Your driverName is set incorrectly, it should be "org.apache.hive.jdbc.HiveDriver".
... View more
06-12-2017
04:04 AM
1 Kudo
Is this a fresh instance? Can you run the following command on the Hive node and post the output? /usr/hdp/current/hive-server2-hive2/bin/schematool -dbType <dbtype> -initSchema -verbose
This will definitely throw more light on what is the issue here.
... View more
06-09-2017
04:38 PM
1 Kudo
There are two problems here with the DDL: Use of delimiter with ORC. ORC is in itself a format so you don't have to provide a delimiter. You would define an external table with a location, also ACID tables cannot be external tables, see here. The proper definition for this table would be: CREATE TABLE FIREWALL_LOGS(
time STRING,
ip STRING,
country STRING,
status INT
)
CLUSTERED BY (time) into 25 buckets
STORED AS ORC
TBLPROPERTIES("transactional"="true");
... View more
06-09-2017
03:55 PM
1 Kudo
The most upto date document to follow for configuring would be this one: https://community.hortonworks.com/articles/101181/rowcolumn-level-security-in-sql-for-apache-spark-2.html
... View more
05-26-2017
08:31 PM
2 Kudos
Hive uses the MySQL jdbc connector for issuing queries to MySQL. You can turn on logging through the jdbc connection url (append &profileSQL=true&logger=com.mysql.jdbc.log.StandardLogger) you are using to connect to mysql db in hive-site.xml. To turn it on the MySQL server you need to set the following in the my.cnf. general_log = 1
general_log_file = /path/to/query.log
See more information here.
... View more
05-24-2017
03:42 PM
1 Kudo
If the goal is to be able to read this data from Hive, then you can use Flume Hive Sink to stream data into Hive table directly.
... View more
05-24-2017
03:27 PM
desc <tablename> shows the table definition (including column names with the datatype) while select * from <tablename> shows the data in the table.
... View more
05-23-2017
11:48 PM
2 Kudos
You can fetch the stats from within hive. See the following thread: https://community.hortonworks.com/questions/103263/cannot-generate-stats-for-partitioned-hive-table.htm You can use the following command: -- In Hive1 for unpartitioned table
SHOW tblproperties <tablename>;
-- In Hive1 for partitioned table
DESC formatted <tablename> partition (<partitionspec>);
-- If using Hive2 on HDP-2.6, following will work on both partitioned and unpartitioned table
DESC formatted <tablename>;
... View more
05-22-2017
05:13 PM
Can you also post namenode logs?
... View more
05-22-2017
05:03 PM
1 Kudo
Did you verify the data by viewing this data using beeline on the cluster? Does it show up correctly lining up under the expected columns when you do a select * from <tablename>. You can run desc [extended|formatted] <tablename> to view the metadata for the table.
... View more
05-19-2017
06:31 PM
1 Kudo
Can you check the following property settings on your cluster? <property>
<name>yarn.scheduler.capacity.queue-mappings-override.enable</name>
<value>false</value>
<description>
If a queue mapping is present and override is set to true, it will override the queue value specified
by the user. This can be used by administrators to place jobs in queues
that are different than the one specified by the user.
The default is false - user can specify to a non-default queue.
</description>
</property> What you desire is true for this parameter.
... View more
05-19-2017
06:53 AM
1 Kudo
I suspect you are hitting a Postgres bug, see https://github.com/pgjdbc/pgjdbc/issues/667. Can you try changing the Postgres JDBC driver to a newer version?
... View more
05-18-2017
08:14 PM
1 Kudo
Hortonworks doesn't provide a .NET plugin but does provide ODBC driver to connect to HiveServer2 service in Hive. Also look at the following thread for more ideas: https://community.hortonworks.com/questions/345/connecting-to-hadoophive-from-net.html
... View more
05-18-2017
07:57 PM
1 Kudo
Overall information is available across partitions only if you are on HDP-2.6 and using Hive2 which i see you are currently not from your description (1.2.1.2.5). If you are using hive 2 the information would be something like: beeline> desc formatted `test_table`;
| Table Parameters: | | |
| | COLUMN_STATS_ACCURATE | {\"BASIC_STATS\":\"true\"} |
| | numFiles | 3 |
| | numRows | 5 |
| | rawDataSize | 21 |
| | totalSize | 26 |
| | transient_lastDdlTime | 1495086467 | In case of Hive 1 unfortunately you will have to run the command per partition and aggregate manually.
... View more
05-18-2017
06:02 PM
2 Kudos
General guidance would be to choose extra large nodes for master nodes and regular for worker nodes. Look at the following articles for detailed guidance: http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.0/bk_cluster-planning/content/ch_hardware-recommendations_chapter.html https://hortonworks.com/blog/deploying-hadoop-cluster-amazon-ec2-hortonworks/ https://community.hortonworks.com/articles/14512/ambari-on-ec2.html You can find non-Ambari manual installation instructions here https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.0/bk_command-line-installation/content/ch_getting_ready_chapter.html
... View more
05-18-2017
06:28 AM
1 Kudo
@Jay SenSharma is right about the logging issue. But that is not the cause of your alert. Basically the metastore alert is failing while running the "show databases" command against the Hive Metastore. You should check whether the server mentioned in the thrift url of hive.metastore.uris is running. Also check the metastore log on that specific host for detailed error. Also on that node check the hive cli log under /tmp/<executing user>/hive.log, where <executing user> might be ambari-qa.
... View more
05-18-2017
06:07 AM
1 Kudo
The stats for partitioned table are available per partition, you can do desc formatted, example: hive> desc formatted `test_table` partition(`date`='2016-12-30');
...
Partition Parameters:
COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
numFiles 1
numRows 2
rawDataSize 9
totalSize 11
transient_lastDdlTime 1495086721
... View more
05-16-2017
07:52 PM
1 Kudo
Hortonworks distribution relies on the Netezza connector provided in the default distribution. SQOOP-846 provided a connector for Netezza in Sqoop-1.4.4.
... View more