Member since
09-25-2015
82
Posts
93
Kudos Received
17
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3849 | 06-06-2017 09:57 AM | |
1066 | 03-01-2017 10:26 AM | |
1083 | 11-22-2016 10:32 AM | |
902 | 08-09-2016 12:05 PM | |
1586 | 08-08-2016 03:57 PM |
08-09-2016
03:53 PM
1 Kudo
@Avijeet Dash
Can you check the value of property is set to "db"- Services-> Ranger>Configs-> Advance->"ranger.audit.source.type" Else you can search for "solr" in filter search tab and check if you have set solr anywhere in config. Let me know if that works.
... View more
08-09-2016
12:05 PM
Do you use Ambari? If so, then there should be no problems as the service user information is recorded in Ambari and propagates through with upgrades. If you don't, then you probably just have to make sure you copy across any relevant configuration parameters, remember to start your services as that user and check HDFS permissions are correct as part of the post-upgrade validation process. As you say, custom users are totally supported, so you shouldn't experience any problems.
... View more
05-12-2016
12:33 PM
@Ana Gillan @Sagar Shimpi Thanks, got partial resolution. Ranger Hive plugin applies only to Hiveserver2 and not to CLI. But in below mentioned hive table file, how user mktg1 is able to query it using HIVE CLI? [hive@sandbox ~]$ hadoop fs -ls /apps/hive/warehouse/xademo.db/customer_details/acct.txt ---------- 3 hive hdfs 1532 2016-03-14 14:52 /apps/hive/warehouse/xademo.db/customer_details/acct.txt [mktg1@sandbox ~]$ hive hive> use xademo;
OK
Time taken: 1.737 seconds hive> select * from customer_details limit 10; OK PHONE_NUM PLAN REC_DATE STAUS BALANCE IMEI REGION
5553947406 6290 20130328 31 0 012565003040464 R06
7622112093 2316 20120625 21 28 359896046017644 R02
5092111043 6389 20120610 21 293 012974008373781 R06
9392254909 4002 20110611 21 178 357004045763373 R04
7783343634 2276 20121214 31 0 354643051707734 R02
5534292073 6389 20120223 31 83 359896040168211 R06
9227087403 4096 20081010 31 35 356927012514661 R04
9226203167 4060 20060527 21 450 010589003666377 R04
9221154050 4107 20100811 31 3 358665019197977 R04
Time taken: 6.467 seconds, Fetched: 10 row(s)
... View more
04-15-2016
03:33 PM
3 Kudos
Hi @krishna sampath, do you want to read the classpath or add to it? To get the whole classpath, you just need to run this command: hadoop classpath If you want to add to it, you can do so in hadoop-env.sh. If you're using Ambari, you must update it in Ambari in the hadoop-env section of the HDFS configs and not on the filesystem directly, as Ambari will overwrite any changes you make to the config files directly.
... View more
09-06-2017
10:58 AM
1 Kudo
Symptom Not able to use reflect function using beeline, but the query works OK with Hvi CLI. Error Message: Error while compiling statement: FAILED: SemanticException UDF reflect is not allowed (state=42000,code=40000) Cause When running set hive.server2.builtin.udf.blacklist from beeline, it will return the following as blacklisted: jdbc:hive2://localhost:10000/default> set hive.server2.builtin.udf.blacklist;
+------------------------------------------------------------------+--+
| set |
+------------------------------------------------------------------+--+
| hive.server2.builtin.udf.blacklist=reflect,reflect2,java_method |
+------------------------------------------------------------------+--+ Reflect UDF is blacklisted by default when running queries through HiveServer2 (beeline, ODBC, JDBC connections), as it was found to be a security risk. The code was modified so if the parameter hive.server2.builtin.udf.blacklist has not been configured or it is blank, its default value will be "reflect,reflect2,java_method". Resolution 1. Open the Ambari UI 2. Add the custom property in Ambari hive.server2.builtin.udf.blacklist under Hive / Configs / Advanced / Custom hive-site and give it any value, for example "empty_blacklist". 3. Restart services as requested by Ambari. 4. Connect again with beeline and verify that blacklist only includes the dummy value now. 0: jdbc:hive2://localhost:10000/default> set hive.server2.builtin.udf.blacklist; +-------------------------------------------+--+
| hive.server2.builtin.udf.blacklist=empty_blacklist |
+-------------------------------------------+--+
5. Reflect should work now without issues.
... View more
03-30-2016
03:04 PM
4 Kudos
@nejm hadj First I’ll answer your question and then I’ll make my recommendation. Answer: The name of the file does not matter. When setting up a Hive external table just specify the data source as the folder that will contain all the files (regardless of names). Details on setting up and external table: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.0/bk_dataintegration/content/moving_data_from_hdfs_to_hive_external_table_method.html Details on reading/parsing JSON files into Hive: http://hortonworks.com/blog/howto-use-hive-to-sqlize-your-own-tweets-part-two-loading-hive-sql-queries/ (alternatively, you can convert JSON to CSV within NiFi. To do so, follow the NiFi portion of this example https://community.hortonworks.com/articles/1282/sample-hdfnifi-flow-to-push-tweets-into-solrbanana.html)
Recommendation: HDFS prefers large files with many entries as opposed to many files with small entries. The main reason being that for each file landed on HDFS, file information is saved in the NameNode (in memory). If you’re putting each twitter message in a separate file you will quickly fill up your NameNodes’s memory and overload the server. I suggest you aggregate multiple messages into one file before writing to HDFS. This can be done with the MergeContent processor in Nifi. Take a look at the below screenshots showing how it would be set up. Also, take a look at the NiFi Twitter_Dashboard.xml example template (https://raw.githubusercontent.com/abajwa-hw/ambari-nifi-service/master/demofiles/Twitter_Dashboard.xml). You can import this into your NiFi by by clicking on Templates (third icon from right) which will launch the 'Nifi Flow templates' popup, and selecting the file.
... View more
01-16-2016
12:23 PM
Thank you very much for your reply and very helpful solutions. I'd rather not manage both a repository HDFS and Hive if I can avoid it. However, we manage Hadoop resources by the YARN queue assigned to each user. For this reason I would like to keep "run as end user instead of hive"(hive.server2.enable.doAs=true).
... View more
12-04-2015
05:15 PM
2 Kudos
@Arpit Agarwal Thanks for helping me out. For others' information: ~50% of dfs.namenode.handler.count is a reasonable value for dfs.namenode.service.handler.count.
... View more
10-22-2015
04:01 PM
JDBC URL specified in Ranger Hive repository configuration is used to provide lookup lists as the users enter database/table/column names in Ranger policy UI. In this scenario, Ranger admin acts as a Hive client to retrieve the names from HiveServer2. The JDBC URL specified should be the same as that of any other client talking to HiveServer2. How do the users access these HiveServer2 instances, if not by ZK discovery?
... View more
- « Previous
-
- 1
- 2
- Next »