Member since
01-19-2017
3681
Posts
633
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1636 | 06-04-2025 11:36 PM | |
| 2087 | 03-23-2025 05:23 AM | |
| 991 | 03-17-2025 10:18 AM | |
| 3772 | 03-05-2025 01:34 PM | |
| 2595 | 03-03-2025 01:09 PM |
04-07-2020
01:42 PM
@SHADA please take note this thread was closed can you open a new thread and attach the errors,logs or screenshot of the error you are encountering and remember to be precise on the version of the sandbox whether it VMware,Docker or Virtualbox Tag me in the new thread
... View more
03-24-2020
03:35 AM
@kasa Can you share a scrambled version of your krb5.conf from both clusters and the auth-to_local of both clusters When copying data from a secure cluster to a secure cluster, the following configuration setting is required in the core-site.xml file: <property>
<name>hadoop.security.auth_to_local</name>
<value></value>
<description>Maps kerberos principals to local user names</description>
</property> Secure-to-Secure: Kerberos Principal Name Assign the same principle name to applicable NameNodes in the source and destination clusters. distcp hdfs://hdp-2.0-secure hdfs://hdp-2.0-secure The SASL RPC client requires that the remote server’s Kerberos principal must match the server principal in its own configuration. Therefore, the same principal name must be assigned to the applicable NameNodes in the source and the destination cluster. For example, if the Kerberos principal name of the NameNode in the source cluster is nn/host1@realm, the Kerberos principal name of the NameNode in destination cluster must be nn/host2@realm, rather than nn2/host2@realm. Secure-to-Secure: ResourceManager mapping rules When copying between two HDP2 secure clusters, further ResourceManager (RM) configuration is required if the two clusters have different realms. Can you share your hadoop.security.auth_to_local on both clusters, in order for DistCP to succeed, the same RM mapping rule must be used in both clusters. I am assuming the REALMS are TEST.COM and DEV.COM for cluster 1 and 2 respectively <property>
<name>hadoop.security.auth_to_local</name>
<value>
RULE:[2:$1@$0](rm@.*CLUSTER1.TEST.COM)s/.*/yarn/
RULE:[2:$1@$0](rm@.*CLUSTER2.DEV.COM)s/.*/yarn/
DEFAULT
</value>
</property> Can you try that and revert
... View more
03-23-2020
04:36 PM
@kasa Disctcp is used for inter/Intracluster copy but the command you are running is not wrong because you need the source and destination NameNodes. $ hadoop distcp /user/home/test.txt /tmp/ The most common use of DistCp is an inter-cluster copy, where you copy from NameNode1[nn1] to Namenode2[nn2] on 2 different clusters and both clusters should be up and running during the process $ hadoop distcp hdfs://nn1:8020/source hdfs://nn2:8020/destination Where hdfs://nn1:8020/source is the data source, and hdfs://nn2:8020/destination is the destination. This will expand the namespace under /source on NameNode "nn1" into a temporary file, partition its contents among a set of map tasks, and start copying from "nn1" to "nn2". Note that DistCp requires absolute paths. Personally I think you should use CopyToLocal instead as according to my understanding you are trying to copy a file from hdfs to you local tmp directory Assuming your directory /user/home/ is in hdfs and you are running the command as HDFS user! This will copy the test.txt from hdfs to local /tmp directory $ hdfs dfs -copyToLocal /user/home/test.txt /tmp/ And to successfully copy between 2 kerberized cluster you should perform the Kerberos cross-realm trust for distcp it's simple to setup just follow the guide and you will be fine Please let me know if my assumption is correct
... View more
03-23-2020
02:55 PM
@Arun66 unfortunately with such a vague and incomplete log, we can't help much. Questions? CDH or HWX Shar e the logs? Share the command being executed? Kerberized or not And any indo you deem important Happy hadooping
... View more
02-03-2020
11:08 AM
@mike_bronson7 I will not have access to my environment for 4 days as I am traveling but I think you can filter using service_name=HDFS I need to test that but that's the way
... View more
02-03-2020
08:36 AM
1 Kudo
@mike_bronson7 Can you run the below it should give the desired response curl -u admin:admin -H "X-Requested-By:ambari" -X GET "http://<ambari-server>:<port>/api/v1/clusters/<clustername>/host_components?HostRoles/stale_configs=false&fields=HostRoles/service_name Please let me know
... View more
01-31-2020
05:58 AM
1 Kudo
@Manoj690 It's always a good idea to share the HDP and Zk version plus the zk logs in /var/log/* having said that can you share your zoo.cfg ? If you really need enable all four letter word commands by default, you can use the asterisk option so you don't have to include every command one by one in the list.See below 4lw.commands.whitelist=* As you have not shared your logs that's a starting point, then restart your zookeeper and let me know!
... View more
01-14-2020
11:43 PM
1 Kudo
@Seaport As the permission is with the zeppelin user [other] you will need to do that at a user level, remember fine-grained security ONLY give what is necessary !! $ hdfs dfs -getfacl /warehouse/tablespace/managed/hive
# file: /warehouse/tablespace/managed/hive
# owner: hive
# group: hadoop
user::rwx
group::---
other::---
default:user::rwx
default:user:hive:rwx
default:group::---
default:mask::rwx
default:other::--- The command below will set [ r-x } bits to the correct ACL you can change to rwx if you wish hdfs dfs -setfacl -R -m user:zeppelin:r-x /warehouse/tablespace/managed/hive Thereafter the zeppelin user can [zeppelin~]$ hdfs dfs -ls /warehouse/tablespace/managed/hive
Found 3 items
drwxrwx---+ - hive hadoop 0 2018-12-12 23:42 /warehouse/tablespace/managed/hive/information_schema.db
drwxrwx---+ - hive hadoop 0 2018-12-12 23:41 /warehouse/tablespace/managed/hive/sys.db
drwxrwx---+ - hive hadoop 0 2020-01-15 00:20 /warehouse/tablespace/managed/hive/zepp.db The earlier error is gone ls: Permission denied: user=zeppelin, access=READ_EXECUTE, inode="/warehouse/tablespace/managed/hive":hive:hadoop:drwx------ Happy hadooping
... View more
01-12-2020
12:56 PM
1 Kudo
@mike_bronson7 When your cluster is in HA it uses a namespace that acts as a load balancer to facilitate the switch from active to standby and vice versa. The dfs-site-xml holds these values filter using dfs.nameservices the nameservice-id should be your namespace or in HA look for dfs.ha.namenodes.[nameservice ID] dfs.ha.namenodes.[nameservice ID] e.g dfs.ha.namenodes.mycluster And that's the value to set eg hdfs://mycluster_namespace/user/ams/hbase The refresh the stale configs , now HBase should sending the metrics to that directory HTH
... View more