Member since
01-19-2017
3681
Posts
633
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1636 | 06-04-2025 11:36 PM | |
| 2088 | 03-23-2025 05:23 AM | |
| 992 | 03-17-2025 10:18 AM | |
| 3772 | 03-05-2025 01:34 PM | |
| 2596 | 03-03-2025 01:09 PM |
07-20-2020
02:39 PM
@focal_fossa AFAIK these sandboxes dynamically allocated storage. You can try that by generate and load data for TPC-DS General usage is tpcds-setup.sh scale_factor [directory] For example below will generate 200 GB of TPC-DS data in /user/data [HDFS] ./tpcds-setup.sh 200 /user/data This should prove that the disk allocation is dynamic see below links https://github.com/hortonworks/hive-testbench/blob/hive14/tpch-build.sh and https://github.com/hortonworks/hive-testbench/blob/hive14/tpch-setup.sh to build Hope that helps
... View more
07-19-2020
09:06 AM
@tanishq1197 The Password should be the scm password Enter SCM password: The default is SCM that should progress successfully
... View more
07-18-2020
01:41 AM
@tanishq1197 I think I did a typo error previously there was an extra scm the below should run successfully sudo /opt/cloudera/cm/schema/scm_prepare_database.sh postgresql scm scm
... View more
07-18-2020
01:18 AM
@tanishq1197 Please can you copy and paste here exactly the command you run?
... View more
07-17-2020
03:24 PM
@Fawze In a Hadoop distribution ACLs are disabled by default. When ACLs are disabled, the NameNode rejects all attempts to set an ACL you will need to enable that manually in CM/Ambari. Enabling HDFS ACLs Using Cloudera Manager Go to the Cloudera Manager Admin Console and navigate to the HDFS service Click the Configuration tab. Select Scope > Service_name (Service-Wide) Select Category > Security Locate the Enable Access Control Lists property and select its checkbox to enable HDFS ACLs. [Enter a Reason for change, and then] Click Save Changes to commit the changes. The above sequence sets the dfs.namenode.acls.enabled property to true in the NameNode's hdfs-site.xml. This is a cluster-wide operation so only after doing the above you can run the setfacl or getfacl commands against an HDFS file etc <property>
<name>dfs.namenode.acls.enabled</name>
<value>true</value>
</property> The hdfs-site.xml is distributed by CM, when you make a change you are usually warned that there is a stale configuration and you are asked to restart the service this triggers the distribution of the new hdfs-site.xml to all nodes in the cluster and that's precisely the reason to use CM rather than manually editing. Happy hadooping
... View more
07-17-2020
01:37 PM
@Fawze AFAIK all changes made manually by editing the files using VI or nano is not persisted in the CM or Ambari database which means CDH or HDP is even not aware of that change so at startup it will get the values of the configuration from the persisted database. When you use CM/Ambari to change values the save button triggers an update in the underlying CM/Ambari tables that said it's NOT advisable to manually edit config files. Happy hadooping
... View more
07-17-2020
11:04 AM
@tanishq1197 I think your syntax is wrong because you have initiated a connection using the below snippet # sudo -u postgres psql The correct syntax is without need of prior log on $ sudo /opt/cloudera/cm/schema/scm_prepare_database.sh [options] <databaseType> <databaseName> <databaseUser> <password> Substituting the correct value s the below should run successfully with the default user/password scm/scm $ sudo /opt/cloudera/cm/schema/scm_prepare_database.sh postgresql scm scm scm Assuming you intend to use the derby databases postgres please see Setting up the Cloudera Manager Database Hope that helps
... View more
07-12-2020
04:58 AM
@Anrygzhang After the merger, the licensing models changed. The last free HDP version that is downloadable is HDP 3.1.4 for any version after that unfortunately would you need to be a Cloudera customer. Get the HDP 3.1.4 link HDP.3.1.4 repository Hope that helps
... View more
07-08-2020
03:04 PM
@chaithanyaam Yes you go into production with Ambari 2.7.4 and HDP 3.1.4 but you will be liable for whatever goes wrong on your production cluster as Cloudera engineers won't be there to rescue you !! Secondly you won't able to upgrade to the newer versions as access to HDP 3.1.5 or higher as these repositories require authentication synonym of buying Cloudera support please check the support options. Even if you have strong technical Hortonworks experts ultimately you will have to move to a paid subscription. For example, HDP 3.1.5 brought in a lot of improvements especially in hive, Hbase, hdfs see fixed issues Having said that I would encourage you also to see the Cloudera Platform pricing Without Cloudera support if anything goes wrong in your cluster you won't get support or any patch or bug fixes from Cloudera unless you buy a subscription I don't think you would want your production cluster running in this mode. Taking into account the above situation it would be wiser to get a subscription for ONLY your production cluster. CDP has for now been released for AWS and Azure I am not sure about the on-premise offering nor a Sandbox Hope that helps
... View more
07-06-2020
08:27 AM
@chaithanyaam AFAIK there is no free version of CDP. You must be a CDP Data Center customer to access these downloads the current CDP release run in AWS and Azure I am not sure when the on-prem offering will be released. Ambari was dropped in favor of Cloudera Manager in CDP but if you really want to continue practicing or working with Cloudera product the best option is using the HDP 3.1.4 which is the last free offering while downloading the HDP 3.1.5 you will need to be a Cloudera customer. The major difference between the HDP 3.1.4 and 3.1.5 is the Hive Warehouse Connector (HWC) Spark and Hive share a catalog in Hive metastore (HMS) instead of using separate catalogs which wasn't the case with earlier versions. The shared catalog simplifies the use of HWC in reading Hive external tables from Spark, you no longer need to define the table redundantly in the Spark catalog. Also, HDP 3.1.5 introduces HMS table transformations. HMS detects the type of client for interacting with HMS, for example, Hive or Spark, and compares the capabilities of the client with the table requirement. Hope that helps
... View more