Member since
01-04-2019
77
Posts
27
Kudos Received
8
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3227 | 02-23-2018 04:32 AM | |
1066 | 02-23-2018 04:15 AM | |
994 | 01-20-2017 02:59 PM | |
1520 | 01-18-2017 05:01 PM | |
4556 | 06-01-2016 01:26 PM |
05-24-2016
09:43 PM
is ranger enabled for Knox? If so make sure you create a ranger knox policy and provide permission for guest user.
... View more
05-11-2016
09:01 PM
You can do two things - create a Hive user defined function that validates each type. or use PIG to do this function prior to loading the data into Hive.
... View more
05-10-2016
08:40 PM
This appears to be FQDN issue. Does your DNS resolution happen through a DNS server or hosts file? if it is hosts file make sure all nodes have fqdn followed by their assigned IP address.
... View more
05-10-2016
08:16 PM
@nfakhar, I had recently created historical build for ETL offload to Hive. We added Effective start date and effective end date to each row. I agree with you adding hash function increases performance for full row compare. The newer version of Hive has Hash functions, but we had used datafu for hashing using PIG.
... View more
05-08-2016
06:33 PM
1 Kudo
Is this a new installation? You should start HBASE from Ambari. HBASE requires zookeeper to be available. Paste the error log for HBASE Master.
... View more
05-08-2016
06:05 PM
Is the datanode service within each slave node appearing started in Ambari?. Also on any one of the slave node that is not working check for errors on /var/log/hadoop/hdfs/<datanodelog> file? Also on slave 1 check namenode log file to see if the datanodes are trying to heartbeat to namenode. Regards Pranay Vyas
... View more
05-08-2016
05:57 PM
1 Kudo
Hi, 1) Number of mappers depends on various factors. primarily number of splits - mapreduce.input.fileinputformat.split.minsize & mapreduce.input.fileinputformat.split.maxsize So a 5GB file configured to have max split size and min split size of 1GB will have 5 mappers. This is just an illustration. See this for Recommended values -> https://community.hortonworks.com/questions/2179/recommended-config-mapreduceinputfileinputformatsp.html 2) Number of containers depends on container size. Read this for calculation of container size http://hortonworks.com/blog/how-to-plan-and-configure-yarn-in-hdp-2-0/ 3) Distcp - read this https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_Sys_Admin_Guides/content/ref-7dbacce5-2629-4e31-b143-e20df092f6d5.1.html Regards Pranay Vyas
... View more
04-07-2016
06:09 PM
Thanks emaxwell and Jason. The problem was due to duplicate HTTP and http account in AD. Deleting the centirfy's 'http' account resolved all issues.
... View more
03-29-2016
08:31 PM
1 Kudo
Hi Jason, 1) Klist from svchdfs says not ticket cache 2) Klist of keytab shows svchdfs-<clustername>@REALM.COM 3) kinit -kt hdfs.headless.keytab svchdfs-<clustername> We noticed that svchdfs-<clustername> exists at 2 OU's within AD. That could be a cause since kerberos is unable to uniquely identify service account. we are trying to delete the duplicate one. Regards Pranay Vyas
... View more
03-29-2016
06:02 AM
1 Kudo
Hello, All services are failing post enabling kerberos with error - "client not found in kerberos database" Kinit yields the same error while using svchdfs account through keytab. kinit to svchdfs works fine if logged in through password. Same error post regenerating keytabs. Appreciate any pointers. 1) HDP 2.3.4.0, Ambari 2.2.0. 2) Pre-created service account are used. 3) AD as Kerberos. 4) AD Structure OU ---level1---> HADOOP ---level1---> cluster1 - serviceprincipals ---level1---> PROD --------level2--------> cluster2 serviceprincipals cluster1 is working fine, cluster2 fails. Regards PranayVyas
... View more
Labels:
- Labels:
-
Apache Ambari