Member since
04-13-2016
422
Posts
150
Kudos Received
55
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1861 | 05-23-2018 05:29 AM | |
4870 | 05-08-2018 03:06 AM | |
1627 | 02-09-2018 02:22 AM | |
2636 | 01-24-2018 08:37 PM | |
6055 | 01-24-2018 05:43 PM |
08-22-2017
07:02 PM
@suresh krish
When you see the environmental variables in your spark UI you can see that particular job will be using below property serialization. If you can't see in cluster configuration, that mean user is invoking at the runtime of the job. <code>spark.serializer org.apache.spark.serializer.KryoSerializer
Secondly spark.kryoserializer.buffer.max is built inside that with default value 64m. If required you can increase that value at the runtime. Even we can all the KryoSerialization values at the cluster level but that's not good practice without knowing proper use case. Hope this helps you.
... View more
07-13-2017
10:39 PM
2 Kudos
If the issue is that the number of rows is too high starting beeline with beeline --incremental=true will be of help
... View more
07-26-2017
06:08 AM
1 Kudo
@Dhiraj Then, you need to run 'Drop database cascade'.
... View more
07-12-2017
03:24 PM
@Sami Ahmad For changing the Ranger Admin password, we need to change the password in 3 different locations. 1. In Ambari you need to change the password at highlighted places: 2. Login to ranger using admin/admin, click on the profile at right-top, then click on change password. Below screenshot helps you to navigate Note: Password should be same across all the places and it should be minimum 8 characters. Hope this helps you.
... View more
10-23-2018
04:44 PM
Its better not to disturb the properties on the statistics usage like hive.compute.query.using.stats. It impacts the way the statistics are used in your query for performance optimization and execution plans. It has tremendous influence on execution plans, the statistics stored depends on the file format as well. Therefore definitely not a solution to change any property with regards to statistics. The real reason for count not working correctly is the statistics not updated in the hive due to which it returns 0. When a table is created first, the statistics is written with no data rows. Thereafter any data append/change happens hive requires to update this statistics in the metadata. Depending on the circumstances hive might not be updating this real time. Therefore running the ANALYZE command recomputes this statistics to make this work correctly.
... View more
06-15-2017
04:56 PM
@vperiasamy Thanks for confirm that Ranger ACL will work as-is. I will debug on it. Thanks for the help, you are the best... 🙂
... View more
07-28-2018
03:29 PM
The script in the blog scans through all the databases on the Hive system, and routes all 'Create' Table statements to a file.. https://tips-to-code.blogspot.com/2018/07/automated-bash-script-to-export-all.html
... View more
06-01-2017
07:59 AM
@spolavarapu thank! It is exactly my case
... View more
05-26-2017
02:07 PM
@Saransh Sharma In addition to @Binu Mathew If you don't want to take separate backups, and there are limited table to take, even you can use Hive Import/Export option. https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport
... View more
06-02-2017
08:10 PM
Thanks @Josh Elser The encoding is done intentionally and hence there is no way to disable/suppress it from REST API Calls, if you wanna do decode it, we need to write Java code accordingly. here is the link: https://hbase.apache.org/book.html#_running_the_shell_in_non_interactive_mode
... View more