Member since 
    
	
		
		
		06-09-2016
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                529
            
            
                Posts
            
        
                129
            
            
                Kudos Received
            
        
                104
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 1732 | 09-11-2019 10:19 AM | |
| 9322 | 11-26-2018 07:04 PM | |
| 2482 | 11-14-2018 12:10 PM | |
| 5311 | 11-14-2018 12:09 PM | |
| 3141 | 11-12-2018 01:19 PM | 
			
    
	
		
		
		12-03-2019
	
		
		11:12 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							Have you tried a path like this -> file:///path/to/mounted/nfs    
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-11-2019
	
		
		10:19 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Seaport   Integration with Ranger is for Atlas Security and in case you want to use tag based policies which is one of the most common use cases for Ranger + Atlas. But is not a requirement, see following link:  https://docs.cloudera.com/HDPDocuments/HDP3/HDP-3.1.0/configuring-atlas/content/additional_requirements_for_atlas_with_ranger_and_kerberos.html     This is also mentioned in our documentation see more in section: Additional Requirements for Atlas with Kerberos without Ranger     Finally, to install Atlas you are only required to have installed Infra Solr, Kafka and Hbase.      Regards,  Felix 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-19-2019
	
		
		09:05 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi @ssankarau      Did you checked if test-user has write permissions on /tmp? I think it may not have permissions to create scratch dir in local. This configuration is set with hive.exec.local.scratchdir.        LOCALSCRATCHDIR("hive.exec.local.scratchdir",
          "${system:java.io.tmpdir}" + File.separator + "${system:user.name}",
          "Local scratch space for Hive jobs"),     By default is set to tmp dit with username, on your case it should default to /tmp/test-user  You can also set this to point somewhere else if necessary.     Hope this helps,  Felix       
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-27-2018
	
		
		12:28 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							  @Danilo Sousa By default ranger comes with an internal admin user (db user). You can reset the password at the DB level back to admin using the following query:  update x_portal_user set password = 'ceb4f32325eda6142bd65215f4c0f371' where login_id = 'admin';  HTH 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-26-2018
	
		
		07:04 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							  @Manikandan Jeyabal Please review this one and let me know if that helps. Perhaps the way the avro is being written is actually causing the problem.   http://mail-archives.apache.org/mod_mbox/avro-user/201105.mbox/%3CCA03B5F3.5891%25Matt.Pouttu-Clarke@icrossing.com%3E  HTH 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-26-2018
	
		
		06:51 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							  @Sura Reddy Have you tried with  // Next change column a1's name to a2, its data type to string, and put it after column b.
ALTER TABLE test_change CHANGE a1 a2 STRING AFTER b;  This is documented as part of https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterTableProperties  HTH  *** If you found this answer addressed your question, please take a moment to login and click the "accept" link on the answer. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-14-2018
	
		
		12:10 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Sriram have you tried configuring as per:  https://community.hortonworks.com/articles/113228/how-to-enable-user-impersonation-for-jdbc-interpre-1.html  Please test the above and let me know if that works for you.  HTH 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-14-2018
	
		
		12:09 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							  @Sriram have you tried configuring as per:  https://community.hortonworks.com/articles/113228/how-to-enable-user-impersonation-for-jdbc-interpre-1.html  Please test the above and let me know if that works for you.  HTH 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-12-2018
	
		
		01:19 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 @sunilprasath elangovan Yes that should be enough, however you may want to configure HS2 with http protocol instead of binary. And for this use-case you may also want to evaluate adding Knox service to the cluster.  HTH 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-05-2018
	
		
		12:25 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Muhammad  Umar   Each executor can have 1 or more threads to perform parallel computation. In yarn master mode the default is 1. These threads can be increased by using the command line parameter --executor-cores.  a) if I have specified 8 num_executors for an application and I dont set executor-cores so will each executor going to use all the cores ?  In yarn master mode the default is 1, therefore each executor will use only 1 core by default.  b) As each node has 8 cores, so what if I specify executor_cores = 4 so that means it will limit executor cores to be used for an executor should not exceed 4 while total cores per node are 8 ?  Assignment of cores is static, not dynamic. And it will remain same during the execution of the application. If you set executor cores to 4 this means that each executor will start and run using 4 cores/threads to perform parallel computation.  c) What is the criteria to specify executor_cores for a spark application?  Increasing executor cores affects performance. You need to take in consideration the number of virtual cores available in each node and my recommendation is you should not increase this over 4 in most cases.  HTH  *** If you found this answer addressed your question, please take a moment to login and click the "accept" link on the answer. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
        













