Member since 
    
	
		
		
		01-04-2019
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                77
            
            
                Posts
            
        
                27
            
            
                Kudos Received
            
        
                8
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 4020 | 02-23-2018 04:32 AM | |
| 1534 | 02-23-2018 04:15 AM | |
| 1372 | 01-20-2017 02:59 PM | |
| 2039 | 01-18-2017 05:01 PM | |
| 5390 | 06-01-2016 01:26 PM | 
			
    
	
		
		
		05-24-2016
	
		
		09:43 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 is ranger enabled for Knox? If so make sure you create a ranger knox policy and provide permission for guest user.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-11-2016
	
		
		09:01 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 You can do two things - create a Hive user defined function that validates each type. or use PIG to do this function prior to loading the data into Hive.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-10-2016
	
		
		08:40 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 This appears to be FQDN issue. Does your DNS resolution happen through a DNS server or hosts file? if it is hosts file make sure all nodes have fqdn followed by their assigned IP address.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-10-2016
	
		
		08:16 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @nfakhar, I had recently created historical build for ETL offload to Hive. We added Effective start date and effective end date to each row. I agree with you adding hash function increases performance for full row compare. The newer version of Hive has Hash functions, but we had used datafu for hashing using PIG. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-08-2016
	
		
		06:33 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Is this a new installation?  You should start HBASE from Ambari. HBASE requires zookeeper to be available.   Paste the error log for HBASE Master.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-08-2016
	
		
		06:05 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Is the datanode service within each slave node appearing started in Ambari?. Also on any one of the slave node that is not working check for errors on /var/log/hadoop/hdfs/<datanodelog> file?  Also on slave 1 check namenode log file to see if the datanodes are trying to heartbeat to namenode.  Regards  Pranay Vyas 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-08-2016
	
		
		05:57 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hi,   1) Number of mappers depends on various factors. primarily number of splits - mapreduce.input.fileinputformat.split.minsize & mapreduce.input.fileinputformat.split.maxsize  So a 5GB file configured to have max split size and min split size of 1GB will have 5 mappers. This is just an illustration.   See this for Recommended values -> https://community.hortonworks.com/questions/2179/recommended-config-mapreduceinputfileinputformatsp.html  2) Number of containers depends on container size. Read this for calculation of container size   http://hortonworks.com/blog/how-to-plan-and-configure-yarn-in-hdp-2-0/  3) Distcp - read this https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_Sys_Admin_Guides/content/ref-7dbacce5-2629-4e31-b143-e20df092f6d5.1.html  Regards  Pranay Vyas 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-07-2016
	
		
		06:09 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Thanks emaxwell and Jason. The problem was due to duplicate HTTP and http account in AD. Deleting the centirfy's 'http' account resolved all issues. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-29-2016
	
		
		08:31 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hi Jason,  1) Klist from svchdfs says not ticket cache  2) Klist of keytab shows svchdfs-<clustername>@REALM.COM  3) kinit -kt hdfs.headless.keytab svchdfs-<clustername>  We noticed that svchdfs-<clustername> exists at 2 OU's within AD. That could be a cause since kerberos is unable to uniquely identify service account. we are trying to delete the duplicate one.   Regards  Pranay Vyas 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-29-2016
	
		
		06:02 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hello,   All services are failing post enabling kerberos with error - "client not found in kerberos database"   Kinit yields the same error while using svchdfs account through keytab. kinit to svchdfs works fine if logged in through password. Same error post regenerating keytabs.   Appreciate any pointers.   1) HDP 2.3.4.0, Ambari 2.2.0.  2) Pre-created service account are used.  3) AD as Kerberos.  4) AD Structure   OU ---level1---> HADOOP        ---level1---> cluster1 - serviceprincipals        ---level1---> PROD       --------level2--------> cluster2 serviceprincipals  cluster1 is working fine, cluster2 fails.   Regards  PranayVyas 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Ambari
 
        













