Member since 
    
	
		
		
		09-24-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                178
            
            
                Posts
            
        
                113
            
            
                Kudos Received
            
        
                28
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 4650 | 05-25-2016 02:39 AM | |
| 4590 | 05-03-2016 01:27 PM | |
| 1195 | 04-26-2016 07:59 PM | |
| 16770 | 03-24-2016 04:10 PM | |
| 3136 | 02-02-2016 11:50 PM | 
			
    
	
		
		
		11-17-2015
	
		
		11:51 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 HDFS restart is required after the change for the new config to take effect permanently. For changing the params as part of the job, this should work -   export HADOOP_OPTS="-Dipc.server.tcpnodelay=true" 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-17-2015
	
		
		06:19 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Is there a way to configure a non-default replication factor for a HDFS directory such that all future files and sub-directories in that directory use that specific replication factor? Currently, we are using a work around of running a daemon process to set the replication factor for all files in the required directory.   Is there a better way to do this?   while true; do
    hdfs dfs -setrep -w 2 /tmp/
    sleep 30
done  I see at one point there was this JIRA https://issues.apache.org/jira/browse/HDFS-199 opened but is blocked by this JIRA [https://issues.apache.org/jira/browse/HADOOP-4771]  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Hadoop
			
    
	
		
		
		11-17-2015
	
		
		05:20 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I don't think this feature is supported natively by Oozie or Falcon. It is a great feature request though. Since Falcon already has the knowledge of DR cluster, it should just replicate the jobs in the secondary cluster. I think the other part of the requirements is that admin should be able to turn on all of the jobs on DR cluster using a single flag / step. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-17-2015
	
		
		04:08 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Is there a recommended rules engine for Hadoop? Has anyone tested Drool with Hive?   The idea is to have a repository of rules, and the engine should read the rules and apply them to the data. These rules could be simple (e.g. value >10) or complex (e.g. average age is 40% greater than the standard deviation of the population). I know that I can write these rules manually in HIVE, but we would like to let the end users make changes on the rules without having us re-write the rule. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Hadoop
- 
						
							
		
			Apache Hive
			
    
	
		
		
		11-02-2015
	
		
		03:39 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							 I know CB creates docker containers with HDP components to provision HDP cluster but can CB also work with a cloud infrastructure that uses Docker as Virtualization layer?  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Docker
- 
						
							
		
			Hortonworks Cloudbreak
			
    
	
		
		
		10-27-2015
	
		
		03:29 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Change the yarn-log4j configuration as shown. Make sure that the properties are uncommented.      
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-26-2015
	
		
		01:31 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 When a new Configuration group is created in Ambari and a newly added host is added to that group, I am unable to edit the settings / properties / configurations for that group. Everything is greyed out.   Verified that admin user is logged in. Also tried creating a brand new configuration group as well as duplicating an existing configuration group but no luck.   What else needs to be done to be able to edit the properties?   I am specifically looking to change the namenode and datanode directories for HDFS. The existing nodes have /grid/[0-3] but the new node only has 1 disk thats mounted to /grid/0 so no [1-3].  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Ambari
			
    
	
		
		
		10-26-2015
	
		
		03:46 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 It is not possible to tell the root cause of the issue by looking at the trace you provided. The log tells that the MR step failed but does not say why. It is possible to get the exact details of the failure though, using the steps below -   1) Click on the job instance in Oozie    2) On the next page, double click on the step that failed.   3) On popup click on the small lens icon button to pull up the log.       4) On the job appilcation page on YARN UI, click on logs and get more information.       In my experience, the error messages are very direct on this page and will tell exactly what the problem is.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-26-2015
	
		
		03:27 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 It could be because the user home directory does not exist on HDFS.  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
         
					
				













