Member since
09-26-2015
48
Posts
29
Kudos Received
6
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
6623 | 10-25-2016 12:53 PM | |
7721 | 10-22-2016 10:22 PM | |
4556 | 10-22-2016 09:34 PM | |
5215 | 10-21-2016 09:56 PM | |
2268 | 07-17-2016 05:26 PM |
10-22-2016
09:46 PM
@S. Sali Do you see this kind of behavior with any other tables or just the bucketed tables ? Can you please check ?
... View more
10-22-2016
09:34 PM
@Riaz Lala Okay, I have got two options for you that you can try: Option#11) Using xpath builtin function: You might have to do some data transformations to your input file to break down the book nodes to individual records. My input data looks like below: <catalog><book><id>11</id><genre>Computer</genre><price>44</price></book></catalog>
<catalog><book><id>45</id><genre>Fantasy</genre><price>5</price></book></catalog>
hive> select * from books_xml;
OK
<catalog><book><id>11</id><genre>Computer</genre><price>44</price></book></catalog>
<catalog><book><id>45</id><genre>Fantasy</genre><price>5</price></book></catalog>
Time taken: 0.483 seconds, Fetched: 2 row(s) Then you can use xpath function like below to access individual columns: hive> select xpath_int(xmldata, '/catalog/book/id'), xpath_string(xmldata, '/catalog/book/genre') as genre, xpath_int(xmldata, '/catalog/book/price') from books_xml;
OK
11 Computer 44
45 Fantasy 5
Time taken: 0.508 seconds, Fetched: 2 row(s) You can do insert overwrite to load into any other external table you want with regular column names Option#2) Use of external UDF function that parses the xml data into an array and use explode function to select the indivudual values and load them to any target table you have. Let me know if this helps
... View more
10-22-2016
08:50 PM
@S. Sali What version of HDP and Hive you are using ? Also can you upload the output of describe formatted grdm.fnl_glbl_currency
... View more
10-22-2016
08:36 PM
@S. Sali Can you post the transcript of the commands you have executed ? Also what version of HDP are you on and what is the Hive version ?
... View more
10-22-2016
08:06 PM
@Riaz Lala Have you looked at the post Let me know if this is helpful
... View more
10-21-2016
09:56 PM
1 Kudo
@Roland Simonis The below documentation has explanation on how to manage permissions they the Ranger --> Settings -->Permissions https://cwiki.apache.org/confluence/display/RANGER/Apache+Ranger+0.5+-+User+Guide#ApacheRanger0.5-UserGuide-PermissionsModule I would say this is a different level of access related grouping from the Ranger operations perspective. To answer your questions: - What does each level of Permissions mean in detail? Permissions tab option allows any user with Admin privileges to control the level of Ranger Admin UI accessibility for Ranger modules to the non-admin users. For example, You can control the Admin access in Ranger if you add/remove users from the Permissions --> Users/Groups tab Resource Based policies option will have a group of users who currently have Resource based policies. You can restrict the access to Resource based policy tab by adding/removing users. If you have Ranger admin access as admin user as well as your user(with admin access), please try logging in as your user user id and try removing yourself and you can see that you wont be able to access the Resource based policies tab after a re-login. You can always add back your access once you log in as admin user. The other options are self explanatory. Let me know if this helps. - How are Users granted permissions automatically? Is there a way to change this, e.g. to stop Ranger from granting all new users permissions on "Resource Based Policies" or "Audit"? Ideally, with User level privileges, you should not be able to see Audit tab, unless admin level access is granted to the user from another admin login. Resource based policies tab will be granted for every user by default, but the users wont be able to see other's policies and they will be able to see only those policies that were exclusively granted for the user thru user level or group level accesses. Again, if you want the users with User level permissions to be restricted to see the Resource Based policy tab at all, you can control that from the Settings --> Permissions --> Resource Based policies option. Let me know if this helps.
... View more
08-04-2016
08:05 PM
@Kumar Veerappan On the Ambari Web UI, click on the "Hosts" tab nd you will find a filter under the Components column using which you can choose what ever component/components you want. You can also filter for the admin host components Thanks Hari
... View more
07-17-2016
05:26 PM
1 Kudo
@AnjiReddy Anumolu Hive transactions are supported starting hive 0.13. However, the integration was completely done for ORC file format for the first release. Look below from the wiki: Only ORC file format is supported in this first release. The feature has been built such that transactions can be used by any storage format that can determine how updates or deletes apply to base records (basically, that has an explicit or implicit row id), but so far the integration work has only been done for ORC. Source: https://cwiki.apache.org/confluence/display/Hive/Hive+Transactions
... View more
07-17-2016
04:48 PM
@Nelson KA Rajendran, Do you mind posting a sample from input file ?How many rows did you have in the input file?
... View more
07-15-2016
02:57 AM
@Peter Kim A1) HDFS block placement by default works in a round-robin fashion across all the directories specified by dfs.data.dir. A2) Its not safe to move one directory contents to another. There is no feature to do intra disk balancing at this moment. The safest way to accomplish this would be to follow the process of de-commissioning the datanode, let Hadoop handle replicating the missing blocks across the cluster, format the drives on the de-commissioned nodes and then recommission back the datanode. Make sure you run the balancer utility in the back end to distribute the blocks evenly across the cluster.
... View more