Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1963 | 07-09-2019 12:53 AM | |
| 11819 | 06-23-2019 08:37 PM | |
| 9103 | 06-18-2019 11:28 PM | |
| 10061 | 05-23-2019 08:46 PM | |
| 4480 | 05-20-2019 01:14 AM |
03-18-2016
03:10 PM
1 Kudo
In HDFS, the permissions model for owner and group follow the BSD rule. The owner is set to the authenticated user, but the group is inherited from the parent directory. This is documented in the Permissions Guide: http://archive.cloudera.com/cdh5/cdh/5/hadoop/hadoop-project-dist/hadoop-hdfs/HdfsPermissionsGuide.html#Overview """ When a file or directory is created, its owner is the user identity of the client process, and its group is the group of the parent directory (the BSD rule). """ The Group Mapping is purely used at the authorisation side, not at the creation side as you are expecting it to be. Since your /user/username directory's group is by default the username itself, that's the value you will naturally see for all groups. If you'd like that changed, you will need to chgrp the /user/username directory to be username:user-group instead of username:username. Subsequent files will now be created with username:user-group under it.
... View more
03-07-2016
04:33 PM
Thanks Harsh. Yes my key was a void .So I changed the avro output to use keyOutputformat (earlier value is now the key) from Avrokeyvalueoutputformat and it worked. Thanks, Nishanth
... View more
03-07-2016
07:37 AM
Hi, I'm sorry for the repeated question but i want to be sure that i cannot use hive acid features with cloudera cdh 5.4.5. Aren't there a way to use the update/delete statement of hive in cdh 5.4.5 or later? Thank you.
... View more
03-03-2016
09:01 AM
Thank you for following up as always, Srini!
... View more
03-03-2016
07:52 AM
One part of HBase Replication is that of turning on configs on the Server end to enable the feature. This can be done via the API exactly like Marcell described. The other part of replication configs is the peer configuration, but this is not doable via CM API as it is not a service-passed configuration, but more of a runtime one. You will need to use HBase's Java API (ReplicationAdmin class) directly for this: http://archive.cloudera.com/cdh5/cdh/5/hbase/apidocs/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.html. To do this in Python, I'd guess you will need to use Jython or such.
... View more
03-03-2016
02:28 AM
Why not creating an Hive table on top of the text file and then simply use an hive query to load the data into the avro table ?
... View more
03-01-2016
02:15 AM
Hello Harsh, Many thanks for your reply ! Got it for the StorageBasedAuthorizationProvider. We already started to use Sentry and has some issues for sharing rights without "Synchronizing HDFS ACLs and Sentry Permissions". The fact we couldn't use impersonation anymore was blocking. I'm going to try the syncronizing feature and everything should go smoothly now. I hope 🙂 Cheers, Pierre
... View more
02-29-2016
01:14 AM
1 Kudo
I would use the spark action as much OOTB as possible leverage sharelib for since handles a number of things for you. You can use multiple versions of sharelib as described here check for overriding the sharelib. Wilfred
... View more
02-28-2016
10:55 PM
1 Kudo
Lets say you want to execute "script.sh" 1. If you have script.sh inside your WF/lib/ path on HDFS, you just need <exec>script.sh</exec> 2. If you have script.sh on an arbitrary path on HDFS, you need: <exec>script.sh</exec> <file>/path/to/script.sh#script.sh</file> 3. Use of the below form with (1) is redundant, but the subsequent form is when you want to invoke it as a different name: <exec>script.sh</exec> <file>script.sh#script.sh</file> <exec>linked-script-name.sh</exec> <file>original-script-name.sh#linked-script-name.sh</file>
... View more
02-28-2016
08:53 AM
1 Kudo
Note: CDH3 is long past its supported lifetime. Netezza JDBC should be worth trying on the CDH3 Sqoop version. I don't recall if it worked without a specialised connector, but the generic SQL connector should likely make it go through.
... View more