Member since 
    
	
		
		
		05-16-2016
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                785
            
            
                Posts
            
        
                114
            
            
                Kudos Received
            
        
                39
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 2327 | 06-12-2019 09:27 AM | |
| 3578 | 05-27-2019 08:29 AM | |
| 5724 | 05-27-2018 08:49 AM | |
| 5242 | 05-05-2018 10:47 PM | |
| 3113 | 05-05-2018 07:32 AM | 
			
    
	
		
		
		03-09-2018
	
		
		04:57 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 For other people reading this in 2018 and beyond NB https://issues.apache.org/jira/browse/HIVE-9452 and https://issues.apache.org/jira/browse/HIVE-17234.  Essentially AFAIK development for an HBase backed metastore has stalled. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-04-2018
	
		
		09:57 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @GeKas  I had tried delete but the limitation is it only supports on ORC file format . (Cloudera says its not production ready) . But in one of our lower enviroment we are using without any issue until now.   Should have mentioned about the limitation that comes with. appreciate your pointer 🙂        
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-27-2018
	
		
		07:07 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi csguna,        Navigator audit is auditing the Hadoop Master roles only, and the hdfs shell commands are working as a regular HDFS client from the NameNode's perspective.     At the namenode side, where HDFS audit logs are generated, is not possible to determine why a client would like to read a file. The only thing that the namenode knows & can log that a client/user would like to open&read a file, but we have no information about what the client will actually do with the data. The client could save the data to a local disk, send it to a network service, simply display the contents of the file, or do an ordinary ETL job and write the results back to HDFS, etc.    That is why an "open" operation is logged for both 'hadoop fs -cat size.log' and 'hadoop fs -get size.log'.  Therefore with Navigator Audit, this is not currently possible, as the knowledge what the client will do with the data read from HDFS is missing.     Usually there are some ways on the OS level itself to audit what users/processes do (like the Linux audit framework), and that can be used to audit file access on the OS level. It might be possible to combine audit data form the OS and Navigator to pinpoint such operations that you mentioned, but I do not know any automated way to do that. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-21-2018
	
		
		10:22 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 as a quick solution, added auto recovery code to our tool, i.e. check datanode role status after cluster startup completion and run start command for each stopped datanode role.        goes like this     Welcome to
     _      __
    | | /| / /______ ____  ___
    | |/ |/ / __/ _ `/ _ \(_-<
    |__/|__/_/  \_,_/ .__/___/   version 1.0.2
                   /_/
starting hdfs-DATANODE-ac7041aa53e984590b7d2e27a66ae6ed
starting hdfs-DATANODE-c3a2e16a1c264acbf2b3a8cd036c8abd
starting hdfs-DATANODE-59269ad5f41a6f45c24d9971f1e45660
starting hdfs-DATANODE-47a0b595206a0616ff011606dff76d0f
waiting 30 sec.
HDFS health checks [0]
+-----------------------------------+---------------+
|               NAME                |    SUMMARY    |
+-----------------------------------+---------------+
|HDFS_BLOCKS_WITH_CORRUPT_REPLICAS  |     GOOD      |
|HDFS_CANARY_HEALTH                 |      BAD      |
|HDFS_DATA_NODES_HEALTHY            |  CONCERNING   |
|HDFS_FAILOVER_CONTROLLERS_HEALTHY  |     GOOD      |
|HDFS_FREE_SPACE_REMAINING          |     GOOD      |
|HDFS_HA_NAMENODE_HEALTH            |     GOOD      |
|HDFS_MISSING_BLOCKS                |     GOOD      |
|HDFS_UNDER_REPLICATED_BLOCKS       |     GOOD      |
+-----------------------------------+---------------+ 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-20-2018
	
		
		10:47 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 No, they reside on the hdfs which resides on the cluster nodes. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-20-2018
	
		
		10:45 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							Something I forgot to mention... this thing was working and all of the sudden stopped.  Port is listening and jar file is in place..  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-20-2018
	
		
		05:07 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @nehanirmal,     If you have a new question regarding Spark, try here:     http://community.cloudera.com/t5/Advanced-Analytics-Apache-Spark/bd-p/Spark       
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-19-2018
	
		
		11:08 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I have already provided the principal and keytab.   UserGroupInformation.loginUserFromKeytabAndReturnUGI(        properties.getProperty("hadoop.kerberos.principal"),  sparkContext.getConf.get("spark.yarn.keytab")  ) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-15-2018
	
		
		12:15 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 ben,  thanks.  the log file is way too big, I just checked again and found the following error:  Can't open /run/cloudera-scm-agent/process/1431-hive-HIVEMETASTORE/supervisor.conf: Permission denied.     all files in directory /run/cloudera-scm-agent/process/   is owned by hive except this file.           -rw-------  1 root root  3430 Feb 15 14:30 supervisor.conf        I have two hive servers, the one works has the same permission issue. but it doesn't prevent it from running successfully.     Here is the more detailed log file information:  + exec /opt/cloudera/parcels/CDH-5.7.6-1.cdh5.7.6.p0.6/lib/hive/bin/hive --config /run/cloudera-scm-agent/process/1431-hive-HIVEMETASTORE --service metastore -p 9083
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/lib/zookeeper/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-5.7.6-1.cdh5.7.6.p0.6/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
18/02/15 14:30:16 ERROR conf.Configuration: error parsing conf core-default.xml
javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown Source)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2541)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2503)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2409)
	at org.apache.hadoop.conf.Configuration.get(Configuration.java:982)
	at org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1032)
	at org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1433)
	at org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:67)
	at org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:81)
	at org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:96)
	at org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool.main(MapreduceDependencyClasspathTool.java:70)
Exception in thread "main" java.lang.RuntimeException: javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2659)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2503)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2409)
	at org.apache.hadoop.conf.Configuration.get(Configuration.java:982)
	at org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1032)
	at org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1433)
	at org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:67)
	at org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:81)
	at org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:96)
	at org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool.main(MapreduceDependencyClasspathTool.java:70)
Caused by: javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown Source)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2541)
	... 9 more
18/02/15 14:30:17 ERROR conf.Configuration: error parsing conf core-default.xml
javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown Source)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2541)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2503)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2409)
	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1144)
	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1116)
	at org.apache.hadoop.mapred.JobConf.setJar(JobConf.java:525)
	at org.apache.hadoop.mapred.JobConf.setJarByClass(JobConf.java:543)
	at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:437)
	at org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2652)
	at org.apache.hadoop.hive.conf.HiveConf.<init>(HiveConf.java:2611)
	at org.apache.hadoop.hive.common.LogUtils.initHiveLog4jCommon(LogUtils.java:74)
	at org.apache.hadoop.hive.common.LogUtils.initHiveLog4j(LogUtils.java:58)
	at org.apache.hadoop.hive.metastore.HiveMetaStore.main(HiveMetaStore.java:6083)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
	at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Exception in thread "main" java.lang.RuntimeException: javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2659)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2503)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2409)
	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1144)
	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1116)
	at org.apache.hadoop.mapred.JobConf.setJar(JobConf.java:525)
	at org.apache.hadoop.mapred.JobConf.setJarByClass(JobConf.java:543)
	at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:437)
	at org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2652)
	at org.apache.hadoop.hive.conf.HiveConf.<init>(HiveConf.java:2611)
	at org.apache.hadoop.hive.common.LogUtils.initHiveLog4jCommon(LogUtils.java:74)
	at org.apache.hadoop.hive.common.LogUtils.initHiveLog4j(LogUtils.java:58)
	at org.apache.hadoop.hive.metastore.HiveMetaStore.main(HiveMetaStore.java:6083)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
	at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: javax.xml.parsers.ParserConfigurationException: Feature 'http://apache.org/xml/features/xinclude' is not recognized.
	at org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown Source)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2541)
	... 18 more
+ date   it looks that the following log files do not have any informationl  /var/log/hive/hadoop-cmf-hive-HIVESERVER2-cahive-master01.log.out and   /var/log/hive/hadoop-cmf-hive-HIVESERVER2-cahive-master01.log.out     also, I tried to start the server from command line and it dosn't work.     $ sudo service hive-server2 start  Redirecting to /bin/systemctl start hive-server2.servic  Failed to start hive-server2.service: Unit not found.        thanks    
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		02-13-2018
	
		
		09:28 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Just quick info you can run pig in local mode as well as in mapreduce mode ,  By default, load looks for your data on HDFS in a tab-delimited file using the default load function PigStorage.    also if you start you pig -x which local mode it will look for local fs .   Nice that you found the fix. @SGeorge , 
						
					
					... View more