Member since 
    
	
		
		
		09-23-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                44
            
            
                Posts
            
        
                49
            
            
                Kudos Received
            
        
                18
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 3112 | 02-08-2017 03:44 PM | |
| 3913 | 02-06-2017 11:03 PM | |
| 4218 | 12-20-2016 02:06 AM | |
| 2023 | 04-29-2016 01:56 PM | |
| 2097 | 04-07-2016 03:11 AM | 
			
    
	
		
		
		05-29-2016
	
		
		05:33 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Can you list the contents of folders "/usr/hdp" and "/usr/hdp/2.4.0.0-169"?   Also the output of "hdp-select | grep zookeeper" on the host where the command failed. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-29-2016
	
		
		01:56 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 It may be https://issues.apache.org/jira/browse/AMBARI-15235 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-07-2016
	
		
		03:11 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		4 Kudos
		
	
				
		
	
		
					
							 As of now, there is no option to do that. You can create few folders as symlinks such as /usr/hdp and /usr/lib/ambari-server, /var/log etc and then install Ambari and the services. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-31-2016
	
		
		04:36 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							 @vpoornalingam can you check if on the host where it is failing you have a "ambari_commons" folder in the /var/lib/ambari-agent/tmp folder? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-30-2016
	
		
		10:02 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							 Thanks for reporting this. This looks like a bug and the work-around you have provided is the valid one till the bug is fixed. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-01-2016
	
		
		07:22 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I believe decommissioned DataNodes are not stopped automatically and they need to be stopped explicitly through API calls. OOTH, decommissioned NodeManagers go down automatically. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		01-05-2016
	
		
		07:09 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Sadly, no. In general, the smaller the blueprint the lesser the chances of breaking compatibility. Let me explain a bit as its not an intention to break backward compatibility. At a very high level, blueprint contains a) components and their layout, b) config settings that are Ambari influenced, c) config settings that are stack influenced, and d) templates for *-env/sh files etc.
In general, 
* (a) remains backward compatible even through major releases.   * (b) and (c) should remain backward compatible unless some config cleanup is performed e.g. created new configs for env/log4j files, cleaned up how heap sizes are specified - rare
* (d) includes scripts - these can go through bug fixes for any release  So if you have a blueprint with primarily (a) and only the configs for which you want to over-write the default it should remain backward compatible. The biggest hurdle you can run into is that when you export a blueprint it does not export a minimal blueprint - rather a blueprint with all sections including (d). 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		01-05-2016
	
		
		03:44 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 The decommission request from does not wait for decommission to complete - it issues the decommission command to hdfs and then hdfs asynchronously continues the decommissioning operation. At this point the jmx data from NN has to be monitored from the outside to know when hdfs completes decommissioning. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		12-30-2015
	
		
		02:14 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 The desired_stack_id for hostcomponentdesiredstate and servicecomponentdesiredstate and current_stack_id for hostcomponentstate should point to HDP-2.3. That could be the reason for the error you see. Also check the version field for hostcomponentstate and see if that is reporting the 2.3 version. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
        











