Member since 
    
	
		
		
		09-18-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                216
            
            
                Posts
            
        
                208
            
            
                Kudos Received
            
        
                49
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 1348 | 09-13-2017 06:04 AM | |
| 2606 | 06-27-2017 06:31 PM | |
| 2499 | 06-27-2017 06:27 PM | |
| 10362 | 11-04-2016 08:02 PM | |
| 9844 | 05-25-2016 03:42 PM | 
			
    
	
		
		
		09-13-2017
	
		
		06:27 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Bhushan kumar What do you mean by ldapsearch not working on Knox? I am assuming that you are trying to do ldapsearch from host where knox is installed and getting "ldapsearch: command not found" which means ldap client utilities are not installed.   Installing openldap-clients should fix the issue.   Below is example from Centos7:  [root@pk-test4 ~]# yum provides ldapsearch
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
 * base: mirror.sjc02.svwh.net
 * extras: mirror.keystealth.org
 * updates: centos.mia.host-engine.com
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo        : base
Matched from:
Filename    : /usr/bin/ldapsearch
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo        : @base
Matched from:
Filename    : /bin/ldapsearch
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo        : @base
Matched from:
Filename    : /usr/bin/ldapsearch
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-13-2017
	
		
		06:04 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 There is no specific Architect certification or exam currently. Below are available certification exams:  https://hortonworks.com/services/training/certification/ 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		07-11-2017
	
		
		07:03 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Muni, yes you can shut down phoenix while HBase will still be running.   On other note, as Tim mentioned below, Pheonix is really good and doesn't consume much resources, try exploring it. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		07-08-2017
	
		
		03:21 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Rishit shah Can you open a new question for this issue? This will help community with better search on issues. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-27-2017
	
		
		06:31 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 @Sami Ahmad Datanode directories is where HDFS data is actually stored while Namenode Directory is used for storing Namenode Metadata i.e. file, block information etc. associated with HDFS data stored on Datanodes. To add additional space to HDFS with additional disks, you should add disk to Datanode Directories i.e. if you add u02 to server(s), add the same to Datanode Directories.   DataNode directories=/u01/hadoop1/hdfs/data,/u02/hadoop1/hdfs/data 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-27-2017
	
		
		06:27 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Well there are many disadvantages of using replication factor 1 and we strongly do not recommend it for below reasons:  1. Data loss --> One or more datanode or disk failure will result in data loss.  2. Performance issues --> Having replication factor of more than 1 results in more parallelization.   3. Handling Failure --> With replication factor > 1, one or more Datanode doesn't result in job failure. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-04-2016
	
		
		08:02 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 It seems you are trying to install Nifi/HDF using Ambari which is managing HDP cluster. At present, HDF cannot be installed using same Ambari which is already managing HDP cluster. You will have to install HDF using dedicated Ambari. Refer below:  https://docs.hortonworks.com/HDPDocuments/HDF2/HDF-2.0.1/bk_ambari-installation/content/ch_getting_ready.html  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-03-2016
	
		
		06:09 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Glad that it got resolved. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-02-2016
	
		
		06:08 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							 To setup an Ambari managed HDP cluster, below services’ components require database for their respective metastore databases.  
 Ambari Server  Hive   Oozie  Ranger   
While a lab/sandbox environment can be setup with default databases for these components, using the same is strongly not recommended for Dev/QA/UAT/Production clusters. Due diligence and planning must be done to ensure that database selection is appropriate for enterprise standard production cluster.   Below are key areas to be taken into consideration while planning to select a database for Ambari and HDP components.  Supported Databases  Refer below for supported databases for Ambari and different HDP components in current state.    https://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-installation/content/database_requirements.html   High Availability Support  It is strongly recommended to setup High Availability for different HDP components, which can be enabled for High Availability in a production cluster.  High Availability is supported by different HDP components as below:    https://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-user-guide/content/ch_managing_service_high_availability.html  The relational database that backs the Hive Metastore, Ambari Server, Oozie Server etc. itself should also be made highly available using best practices defined for the database system in use for HDP services and Ambari Server to be truly Highly Available and not to have database as single point of failure for the service.  Therefore, it is important to select relational database, which supports highly availability, and it should be discussed with in-house DBA when planning for a new database or to use an existing in-house database for HDP deployment.  Cost of licensing and support  HDP support subscription doesn’t cover any licensing and support for databases being used for Ambari Server and HDP components i.e. Hive Metastore and would incur additional licensing and support cost. Therefore, cost of licensing and support should be considered as an important factor for selecting appropriate database for Ambari Server and HDP stack.  Note: Contact in-house database team/database vendor for details on cost for licensing and support for databases.  Database maintenance and management  Database to be used for Ambari Server and HDP components would need maintenance and management which can be quite frequent/regular for database backup, HA setup and recovery etc. Therefore, while selecting a database for Ambari/HDP, it must be ensured that your organization has in-house skilled people/DBAs available to perform these activities.  It is not a good practice to use different relational databases for different components i.e. Postgres for Ambari, MySQL for Hive etc. to avoid complexity for management and maintenance of these different databases. It is recommended to pick relational database of your choice and use the same i.e. MySQL for all components or Postgres for all components and so on. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
         
					
				













