Member since
09-18-2015
216
Posts
208
Kudos Received
49
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1037 | 09-13-2017 06:04 AM | |
2085 | 06-27-2017 06:31 PM | |
2005 | 06-27-2017 06:27 PM | |
8830 | 11-04-2016 08:02 PM | |
9168 | 05-25-2016 03:42 PM |
09-13-2017
06:27 AM
@Bhushan kumar What do you mean by ldapsearch not working on Knox? I am assuming that you are trying to do ldapsearch from host where knox is installed and getting "ldapsearch: command not found" which means ldap client utilities are not installed. Installing openldap-clients should fix the issue. Below is example from Centos7: [root@pk-test4 ~]# yum provides ldapsearch
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
* base: mirror.sjc02.svwh.net
* extras: mirror.keystealth.org
* updates: centos.mia.host-engine.com
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo : base
Matched from:
Filename : /usr/bin/ldapsearch
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo : @base
Matched from:
Filename : /bin/ldapsearch
openldap-clients-2.4.40-13.el7.x86_64 : LDAP client utilities
Repo : @base
Matched from:
Filename : /usr/bin/ldapsearch
... View more
09-13-2017
06:04 AM
There is no specific Architect certification or exam currently. Below are available certification exams: https://hortonworks.com/services/training/certification/
... View more
07-11-2017
07:03 PM
Muni, yes you can shut down phoenix while HBase will still be running. On other note, as Tim mentioned below, Pheonix is really good and doesn't consume much resources, try exploring it.
... View more
07-08-2017
03:21 AM
@Rishit shah Can you open a new question for this issue? This will help community with better search on issues.
... View more
06-27-2017
06:31 PM
1 Kudo
@Sami Ahmad Datanode directories is where HDFS data is actually stored while Namenode Directory is used for storing Namenode Metadata i.e. file, block information etc. associated with HDFS data stored on Datanodes. To add additional space to HDFS with additional disks, you should add disk to Datanode Directories i.e. if you add u02 to server(s), add the same to Datanode Directories. DataNode directories=/u01/hadoop1/hdfs/data,/u02/hadoop1/hdfs/data
... View more
06-27-2017
06:27 PM
1 Kudo
Well there are many disadvantages of using replication factor 1 and we strongly do not recommend it for below reasons: 1. Data loss --> One or more datanode or disk failure will result in data loss. 2. Performance issues --> Having replication factor of more than 1 results in more parallelization. 3. Handling Failure --> With replication factor > 1, one or more Datanode doesn't result in job failure.
... View more
11-04-2016
08:02 PM
1 Kudo
It seems you are trying to install Nifi/HDF using Ambari which is managing HDP cluster. At present, HDF cannot be installed using same Ambari which is already managing HDP cluster. You will have to install HDF using dedicated Ambari. Refer below: https://docs.hortonworks.com/HDPDocuments/HDF2/HDF-2.0.1/bk_ambari-installation/content/ch_getting_ready.html
... View more
11-03-2016
06:09 PM
Glad that it got resolved.
... View more
11-02-2016
06:08 PM
2 Kudos
To setup an Ambari managed HDP cluster, below services’ components require database for their respective metastore databases.
Ambari Server Hive Oozie Ranger
While a lab/sandbox environment can be setup with default databases for these components, using the same is strongly not recommended for Dev/QA/UAT/Production clusters. Due diligence and planning must be done to ensure that database selection is appropriate for enterprise standard production cluster. Below are key areas to be taken into consideration while planning to select a database for Ambari and HDP components. Supported Databases Refer below for supported databases for Ambari and different HDP components in current state. https://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-installation/content/database_requirements.html High Availability Support It is strongly recommended to setup High Availability for different HDP components, which can be enabled for High Availability in a production cluster. High Availability is supported by different HDP components as below: https://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-user-guide/content/ch_managing_service_high_availability.html The relational database that backs the Hive Metastore, Ambari Server, Oozie Server etc. itself should also be made highly available using best practices defined for the database system in use for HDP services and Ambari Server to be truly Highly Available and not to have database as single point of failure for the service. Therefore, it is important to select relational database, which supports highly availability, and it should be discussed with in-house DBA when planning for a new database or to use an existing in-house database for HDP deployment. Cost of licensing and support HDP support subscription doesn’t cover any licensing and support for databases being used for Ambari Server and HDP components i.e. Hive Metastore and would incur additional licensing and support cost. Therefore, cost of licensing and support should be considered as an important factor for selecting appropriate database for Ambari Server and HDP stack. Note: Contact in-house database team/database vendor for details on cost for licensing and support for databases. Database maintenance and management Database to be used for Ambari Server and HDP components would need maintenance and management which can be quite frequent/regular for database backup, HA setup and recovery etc. Therefore, while selecting a database for Ambari/HDP, it must be ensured that your organization has in-house skilled people/DBAs available to perform these activities. It is not a good practice to use different relational databases for different components i.e. Postgres for Ambari, MySQL for Hive etc. to avoid complexity for management and maintenance of these different databases. It is recommended to pick relational database of your choice and use the same i.e. MySQL for all components or Postgres for all components and so on.
... View more