Member since
05-16-2016
785
Posts
114
Kudos Received
39
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1834 | 06-12-2019 09:27 AM | |
3046 | 05-27-2019 08:29 AM | |
5073 | 05-27-2018 08:49 AM | |
4448 | 05-05-2018 10:47 PM | |
2763 | 05-05-2018 07:32 AM |
10-13-2021
11:11 PM
How did you resolve the issue? I am facing a similar problem with datanode not starting
... View more
05-24-2021
06:46 PM
Hi below are some pointers just in case not tried ( which I believe you using them ) set the column value to TIMESTAMP -- ( map-column-hive <cols_name>=TIMESTAMP) Then please keep in mind the column should be bigint The main issue is the format. Parquet represent time in "msec" where as impala will interpret BigInt as "sec" so to get correct value you need to take care at query level. ( divide by 1000) Regards Jay
... View more
03-10-2021
12:00 AM
@Venkat_ as this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post.
... View more
02-25-2021
02:56 AM
Please mark and give us kudos so it encourages the fellow users to contribute more !
... View more
02-18-2021
11:32 PM
Hi, I got the same error in hue 3.1.0. The sql works fine with beeline but will get same error with hue. And we're also using impala hive HA. Tried [Set use_get_log_api=true in the beeswax section ], and it's not working. Do you have any other solution?
... View more
12-11-2020
05:50 AM
The titled SocketTimeoutException occurs when the thrift-client in hiveConnection object is in the process of actively reading sql results from hive server2 (thrift server), and is not able to receive anything until the TSocket's time out occurs. You can check the source code from:HiveConnection.setupLoginTimeout and HiveAuthFactory.getSocketTransport. So you need to either tuning hiveserver2, or increase the TSocket's timeout setting. And for now, the only way to increase Tsocket's time out setting is via: DriverManager.setLoginTimeout() you can check below jira for more information: https://issues.apache.org/jira/browse/HIVE-22196 https://issues.apache.org/jira/browse/HIVE-6679 https://issues.apache.org/jira/browse/HIVE-12371
... View more
07-16-2020
05:29 PM
When you add the node, there is a script called allkeys.sh will generate a key bundle, it contains the GPG key info and key bundle is called allkeys.asc DEFAULT_CLOUDERA_KEY_BUNDLE_NAME = "allkeys.asc" The key bundle will get the key for each of the flavor, so in your case it is archive.key which is located here : https://archive.cloudera.com/cdh5/ubuntu/lucid/amd64/cdh/archive.key If it is RHEL, then it uses this: https://archive.cloudera.com/cdh5/redhat/7/x86_64/cdh/RPM-GPG-KEY-cloudera Once all these keys are downloaded, it will be signed by the master key. Finally gpg command is used to export the keys to bundle called allkeys gpg --export -a > allkeys So I would check what is the repo that is being used
... View more
06-27-2020
12:40 PM
Is ther any solution for this issue. I am facing same issue. Table is holding very huge data and while doing insert overwrite , files are getting placed in my user directory, /user/anjali/.Trash, causing hive action in oozie to fail after 1.5 hr long run. Please help. The table is external and ev even I changed it to internal table, auto purge = true is not working.
... View more
03-03-2020
09:30 PM
1 Kudo
Could you run the administrator cmd not a normal one? From your error, I think you have not enough privilege to run these commands.
... View more