Member since 
    
	
		
		
		09-24-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                816
            
            
                Posts
            
        
                488
            
            
                Kudos Received
            
        
                189
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 3111 | 12-25-2018 10:42 PM | |
| 13999 | 10-09-2018 03:52 AM | |
| 4690 | 02-23-2018 11:46 PM | |
| 2405 | 09-02-2017 01:49 AM | |
| 2826 | 06-21-2017 12:06 AM | 
			
    
	
		
		
		06-22-2025
	
		
		11:03 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hello @sunile_manjee Sir,  I have used above commands and I am able to export the data into csv, But it the data is enclosed in quotes, (mentioned below). Is there a way to get the data without 'quotes' and also the generated file has Header, is there a way to generate the file without Header.  Appreciate your help. Thanks     'Column1', 'Column2'  'Data1', 'Data2' 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-08-2022
	
		
		02:45 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hello @hbasetest You wish to enable Normalizer at Cluster Level irrespective of the Table Level Setting i.e. NORMALIZATION_ENABLED be True or False. As far as I believe, We would require Table Level enabling.      Having said that, If you can Open a Post on the same by using the Steps shared by @VidyaSargur, Our fellow Community Gurus can get back to you sooner, as compared to a Comment on an Article written in 2016.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-27-2021
	
		
		11:15 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Venk123 as this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-14-2021
	
		
		08:13 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 A bit late to the party, but hope the following will help.     By calling the main functions of the classes,   UnixUserGroupBuilder, PolicyMgrUserGroupBuilder or LdapUserGroupBuilder  is not going to work, since the main classes of these are only initializing the classes. In order to start the actual sync, the function updateSink needs to be called. During startup this is handled by the class  org.apache.ranger.usergroupsync.UserGroupSync  thus, calling its main function will trigger the syncing using the configuration that you set in your cluster.     A complete example for triggering the usersync manually could be:  java -Dlogdir=/var/log/ranger/usersync -cp "/usr/hdp/current/ranger-usersync/dist/unixusersync-1.2.0.3.1.5.135-2.jar:/usr/hdp/current/ranger-usersync/lib/*:/etc/ranger/usersync/conf" org.apache.ranger.usergroupsync.UserGroupSync  for HDP and  java -Dlogdir=/var/log/ranger/usersync -cp "/opt/cloudera/parcels/CDH/lib/ranger-usersync/dist/unixusersync-2.1.7.1.7.0-460.jar:/opt/cloudera/parcels/CDH/lib/ranger-usersync/lib/*:/etc/ranger/usersync/conf" org.apache.ranger.usergroupsync.UserGroupSync  for CDP. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-21-2021
	
		
		12:25 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi @Rajuambala as this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		01-04-2021
	
		
		08:30 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @kalhan      While it is possible to have a single ZK cluster to support multiple services, It is the recommendation that NiFi have its own dedicated ZK cluster.  NiFi cluster stability is dependent on ZK and many of the NiFi processors that can be used depend on on Cluster state which is also stored in ZK.  IF ZK becomes overburdened it can affect overall stability and performance of NiFi.    If you found any of the answers provided on this query helped you, please select "accept solution" on each of them.    Thank you,  Matt    Hope this helps. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		12-04-2020
	
		
		01:27 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @sshimpi I am getting this failure error when I do gradle clean build :        The 'sonar-runner' plugin has been deprecated and is scheduled to be removed in Gradle 3.0. please use the official plugin from SonarQube (http://docs.sonarqube.org/display/SONAR/Analyzing+with+Gradle). :clean :buildInfo :compileJava FAILURE: Build failed with an exception. * What went wrong: Could not resolve all dependencies for configuration ':compile'. > Could not resolve com.sequenceiq:ambari-client20:2.0.1. Required by: com.sequenceiq:ambari-shell:0.1.DEV > Could not resolve com.sequenceiq:ambari-client20:2.0.1. > Could not get resource 'http://maven.sequenceiq.com/snapshots/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > Could not GET 'http://maven.sequenceiq.com/snapshots/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > maven.sequenceiq.com: Name or service not known > Could not resolve com.sequenceiq:ambari-client20:2.0.1. > Could not get resource 'http://maven.sequenceiq.com/release/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > Could not GET 'http://maven.sequenceiq.com/release/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > maven.sequenceiq.com > Could not resolve com.sequenceiq:ambari-client20:2.0.1. > Could not get resource 'http://maven.sequenceiq.com/releases/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > Could not GET 'http://maven.sequenceiq.com/releases/com/sequenceiq/ambari-client20/2.0.1/ambari-client20-2.0.1.pom'. > maven.sequenceiq.com * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. BUILD FAILED Total time: 11.708 secs        Please help! 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		12-01-2020
	
		
		03:47 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I'm trying to run a dag with airflow 1.10.12 and HDP 3.0.0  when i run the dag it gets stuck in ```Connecting to jdbc:hive2://[Server2_FQDN]:2181,[Server1_FQDN]:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2```  when i run ```beeline -u "jdbc:hive2://[Server1_FQDN]:2181,[Server2_FQDN]:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2"``` from shell, it connect to hive with no problem.  I've also made a connection like this  ```  Conn Id *  hive_jdbc  -------------  Conn Type  -------------  Connection URL  jdbc:hive2://centosserver.son.ir:2181,centosclient.son.ir:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2  -------------  Login  hive  -------------  Password  ******  -------------  Driver Path  /usr/hdp/3.0.0.0-1634/hive/jdbc/hive-jdbc-3.1.0.3.0.0.0-1634-standalone.jar  -------------  Driver Class  org.apache.hive.jdbc.HiveDriver  ```  and I'm not using kerberos  I've also added ```hive.security.authorization.sqlstd.confwhitelist.append``` in the ambari ```Custom hive-site```  ```  radoop\.operation\.id|mapred\.job\.name||airflow\.ctx\.dag_id|airflow\.ctx\.task_id|airflow\.ctx\.execution_date|airflow\.ctx\.dag_run_id|airflow\.ctx\.dag_owner|airflow\.ctx\.dag_email|hive\.warehouse\.subdir\.inherit\.perms|hive\.exec\.max\.dynamic\.partitions|hive\.exec\.max\.dynamic\.partitions\.pernode|spark\.app\.name  ```  any suggestions? I'm desperate, I've tried every way that i know but still nothing  @nsabharwal @agillan @msumbul1 @deepesh1  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-11-2020
	
		
		01:40 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I did this in root user, found the file and changed it there. But, how to change it for each node?    
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-29-2020
	
		
		10:11 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi,      has anyone get this running and can post an running example ?      thx  marcel        
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
         
					
				













