Member since
01-16-2018
541
Posts
33
Kudos Received
82
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
114 | 01-18-2023 12:10 AM | |
75 | 01-16-2023 01:54 AM | |
172 | 01-13-2023 01:59 AM | |
163 | 01-13-2023 01:35 AM | |
100 | 01-02-2023 10:03 PM |
10-14-2022
03:31 AM
Greetings @Khairul_Hasan Hope you are doing well. We wish to follow-up on the above Post. Kindly let us know the Outcome of the above ask from our side. Note that in 7.1.3, CDP uses Phoenix 5.0 & in 7.1.7, CDP have Phoenix 5.1. Henceforth, our ask to include the Phoenix Server & Client Jar explicitly. Regards, Smarak
... View more
10-14-2022
03:26 AM
Hello @SDL This is an Old Thread & I assume your Team have moved on, yet wish to Update this Post for future references. It was observed that such Overnight Restart were resetting the default CleanUp (24 Hours) set via [1] in SolrConfig.XML of the respective Solr Collection (Sample from Ranger_Audits Collection). This caused the CleanUp to be postponed on a daily basis & causes Document PileUp beyond their Expiration. If Customer are restarting the Service nightly, It's advisable to set the CleanUp from 24 Hours to a Lower Value (Like, 20 or 22 Hours). Regards, Smarak [1] <processor class="solr.processor.DocExpirationUpdateProcessorFactory"> <int name="autoDeletePeriodSeconds">86400</int> <str name="ttlFieldName">_ttl_</str> <str name="expirationFieldName">_expire_at_</str> </processor>
... View more
10-12-2022
03:16 AM
Hello @utehrani Thanks for engaging Cloudera Community. Based on the Post, Pods for CML on ECS are failing, which isn't being resolved by Restarting the Pod as well. The Error [1] is generally observed for few Checks/Configs failing & frankly, the review of such an issue wouldn't be easy with Community. Since your Team installed ECS & deploying CML, We believe it's quicker for your Team to engage Cloudera Support via a Support Case while sharing the Version (CDP Private Cloud Base Version, Private Cloud Data Service Version) along with the Pod Listing "kubectl get pods -A" & the Logs from the affected Pods. Also, Whether your Team have deployed other Data Services like CDW (Cloudera Data Warehouse) & CDE (Cloudera Data Engineering) on the concerned ECS Setup yet. If No, We would like to confirm if their Deployment is working fine or facing similar concerns. Regards, Smarak [1] Failed to obtain initialization data due to {} com.cloudera.cdp.CdpServiceException: com.cloudera.cdp.CdpServiceException: 404: NOT_FOUND
... View more
10-10-2022
09:27 PM
Hello @cprakash Since we haven't heard from your Team, We are marking the Post as Resolved. Feel free to add your Team's observation whenever feasible. In Summary, Review the HMaster Logs to confirm the reasoning for ConnectionRefused. Few possible scenarios being Port 16000 is being used by any other Service Or, "master1" isn't correctly being mapped as per DNS Or, Port 16000 may be blocked. Regards, Smarak
... View more
10-10-2022
09:21 PM
Hello @Kings Thanks for using Cloudera Community. Based on your Post, You wish to reuse the Hardware used for MapR to run Cloudera with the Data Storage Setup. The Hardware/Java/Network/Database requirement for CDP Private Cloud v7.1.8 (Latest Release) is available via [1]. Your Team can review the same & confirm if the existing Hardware & Setup meets the requirement. Note that your Team needs to be install Cloudera Manager & allow Cloudera Manager to perform the Installation of various Services (Including HDFS) across the Cluster Hosts. Whether a Complete Lift & Shift works depends on the existing Setup & compliance with [1]. For information, Cloudera offers a Public Cloud & SaaS offering as well. Additionally, Customer can use the existing Hardware to turn into a full-fledged Kubernetes Cluster via Cloudera Manager to run Kubernetes based Workload (CDE, CDW, CML). The reason for stating the same is your Team selected "CDE" i.e. Cloudera Data Engineering in the Labels. We believe this is a right point to engage with our Cloudera's Sales Team & find the exact fit for your Team's requirement in Cloudera's Hybrid Cloud Offering. Kindly review & let us know. Regards, Smarak [1] https://docs.cloudera.com/cdp-private-cloud-base/7.1.8/installation/topics/cdpdc-hardware-requirements.html
... View more
10-10-2022
09:01 PM
Hello @Khairul_Hasan Thanks for engaging Cloudera Community. Based on the Post, Your Team is receiving [1] while running the Command [2] in CDP v7.1.7. The Same Command was running with CDP v7.1.3. If you include Verbose for ClassLoading (Similar to "-verbose:class"), Your Team would confirm the Jar from which the Method is loaded. If your Team can pass the Phoenix Client & Phoenix Server jar explicitly in the ClassPath (The Path would be similar to "/opt/cloudera/parcels/<PhoenixParcelDir>/lib/phoenix/<PhoenixClientJar>" & "/opt/cloudera/parcels/<PhoenixParcelDir>/lib/phoenix/<PhoenixServerJar>"), We expect the Error to be managed. Kindly review the same & share the Outcome. Regards, Smarak [1] Can't find method newStub in org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService! [2] java -cp /etc/hadoop/conf.cloudera.hdfs/ssl-client.xml:/etc/hbase/conf.cloudera.hbase/hbase-site.xml:/etc/hadoop/conf.cloudera.hdfs/core-site.xml:/etc/hadoop/conf.cloudera.hdfs/hdfs-site.xm:/data/scripts/LeaApp-1.0-SNAPSHOT.jar net.ba.lea.transformation.FileActions "/data/scripts/msc/IN/" "/tmp/nss_processing/" "/data/scripts/msc/reject/" "250" "LEA.DBM_CDR_FILE_HEAD" "NSS" "jdbc:phoenix:gzvlcdpnode01.ba.net:2181:/hbase:phoenix/gzvlcdpnode02@BA.NET:/etc/security/keytab/phoenix.keytab"
... View more
08-17-2022
01:09 AM
Hello All, This is an older post which had a few recent followup queries. To close the loop, HBase offers multiple Tools to migrate Data from 1 Cluster to another Cluster like Snapshot, Export-Import, HashTable/SyncTable etc. Most of these Tools relies on MapReduce & uses 1 Mapper per Region of the Source Table. All these Tools works without any concerns. The only part of the ask which can't be answered accurately is the Concurrency/Job Configurations/Mapper Memory etc. These details rely on Customer's Environment Setup & the Bandwidth between the 2 Clusters. As such, Customer can run 1 such HBase MR Job & see the Outcome. Accordingly, Fine-Tune is required. If any issues are observed while performing the above HBase MR Job, Feel free to post the Q in a Community Post for fellow Community Members to review & share their thoughts. Regards, Smarak
... View more
08-08-2022
02:45 AM
Hello @hbasetest You wish to enable Normalizer at Cluster Level irrespective of the Table Level Setting i.e. NORMALIZATION_ENABLED be True or False. As far as I believe, We would require Table Level enabling. Having said that, If you can Open a Post on the same by using the Steps shared by @VidyaSargur, Our fellow Community Gurus can get back to you sooner, as compared to a Comment on an Article written in 2016.
... View more
08-04-2022
02:28 AM
Hello @achandra, This is an Old Post yet closing the same by sharing the feedback concerning your ask for wider audience. The API is failing owing to Space between "NOW-" & "7DAYS". There shouldn't be any gap between the same. In Summary, the Command is below, where Customer needs to set the HTTP(s) header, Solr Host & Solr Port accordingly. Additionally, the Example uses "ranger_audits" Collection & "evtTime" field to delete any Documents older than 7 Days: ### curl -k --negotiate -u : "http[s]://<Any Solr Host FQDN>:<Solr Port>/solr/ranger_audits/update?commit=true" -H "Content-Type: text/xml" --data-binary "<delete><query>evtTime:[* TO NOW-7DAYS]</query></delete>" Regards, Smarak
... View more
07-06-2022
07:49 AM
Hello @LakshmiSegu We hope your Q concerning the Zeppelin Access issue is addressed by our 06/21 Post. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
07-06-2022
07:47 AM
1 Kudo
Hi @cprakash Kindly review our Post on 06/21 & share your observation. If you have resolved the issue, Kindly share the Solution for our wider Community Members. Regards, Smarak
... View more
07-06-2022
07:41 AM
Hello @sgenchev We hope your Q concerning identifying the Region/RegionServers was addressed by our 06/22 Post. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
07-06-2022
07:40 AM
Hello @caisch We hope our Post on 06/21 help answer your query concerning Solr TTL. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
06-22-2022
12:15 AM
Hello @sgenchev While I am not familiar with building Charts to answer the Query, I shall share a few CLI approaches: (I) HBCK: Running "hbase hbck --details" prints [1] detailing the Regions per RegionServer. (II) Using JMX: Say, the HMaster UI is "http://HMaster1:22002/master-status". Replace "master-status" with "jmx", which prints the JSON Output. Additional filtering can be done to capture the Regions per RegionServers. Regards, Smarak [1] HBCK Output: RegionServer: RegionServer1,22101,1655811831192 number of regions: 1 RegionServer: RegionServer2,22101,1655811830393 number of regions: 1 RegionServer: RegionServer2,22101,1655811832307 number of regions: 1
... View more
06-21-2022
11:43 PM
Hello @cprakash Let us know if you have fixed the ConnectionRefused Error to Master1 on Port 16000. As @rki_ mentioned, Review the HMaster Logs to confirm the reasoning for ConnectionRefused. Few possible scenarios being Port 16000 is being used by any other Service Or, "master1" isn't correctly being mapped as per DNS Or, Port 16000 may be blocked. Based on the HMaster Logs, We may offer further review. As such, If the Issue persists, Share the HMaster Logs covering the concerned Stacktrace as well. If the Issue has been resolved, We shall appreciate if you can share the details into the Cause & Fix applied. Regards, Smarak
... View more
06-21-2022
11:38 PM
Hello @LakshmiSegu We hope your query was addressed by Shehbaz's response. In Summary, (I) Ensure you Username has an IDBroker Mapping (Actions > Manage Access > IDBroker Mappings). (II) Include the "spark.yarn.access.hadoopFileSystems" Parameter to point to the S3 Path [1]. Regards, Smarak [1] https://docs.cloudera.com/runtime/7.2.15/developing-spark-applications/topics/spark-s3.html
... View more
06-21-2022
11:17 PM
Hello @caisch Thanks for using Cloudera Community. Based on your Post, you wish to confirm if TTL for HDFS can be set to 90Days & other Services' TTL is set to 14Days. Since you selected Solr, Let me answer your Post & you can let me know if I understood the Post differently. In Solr, We have Collections which may have Data on HDFS or Local. For Example, RangerAudits Collection may have Data on HDFS & Atlas Collection may have Data on Local. At each Collection Level, the "solrconfig.xml" captures the TTL [1] via DocExpirationUpdateProcessorFactory Class. You can configure the TTL at each Collection Level in Solr & they would Cleanup the Underlying Data, be it on HDFS or Local. Using the above Example of RangerAudits using HDFS Storage & Atlas using Local Storage, We can set RangerAudits to expire at 90 Days & Atlas to expire at 14 Days, which in turn would remove the Underlying Data from HDFS & Local respectively for RangerAudits & Atlas. Kindly review & let us know if the above answers your Post. If No, You may clarify & we shall get back to you accordingly. Regards, Smarak [1] https://solr.apache.org/docs/8_4_0/solr-core/org/apache/solr/update/processor/DocExpirationUpdateProcessorFactory.html
... View more
04-19-2022
01:07 AM
Hello @SVK If your queries concerning Apache Airflow has been addressed, Feel free to mark the Post as Solved. If you have any further ask, Kindly share the same & we shall get back to you accordingly. Regards, Smarak
... View more
04-19-2022
01:05 AM
Hello @HiThere We hope your query concerning VMs & recommendations around Hardware is answered. We are marking the Post as Closed. If you have any further concerns, Feel free to post your ask & we shall answer your queries. Regards, Smarak
... View more
04-14-2022
01:29 AM
Hello @HiThere Thanks for using Cloudera Community. To your Query, Kindly refer [1] & [2] for Resource Requirement for CDP v7.1.7. Note that Documentation refers to Hardware requirement in terms of Resources (CPU, Memory, Network, Disk) as opposed to Physical Machines or Virtual Machines. As long as your Team meet the Hardware requirement for Storage & Compute, the Virtualized & Bare-Metal choices shouldn't matter. Regards, Smarak [1] https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/installation/topics/cdpdc-hardware-requirements.html [2] https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/concepts/topics/cm-vpc-networking.html
... View more
04-14-2022
01:23 AM
Hello @SVK Thanks for using Cloudera Community. Based on your post, You wish to confirm if Airflow is Supported by Cloudera. Summarizing the responses shared by my Colleagues for you: (I) Airflow isn't Supported by Cloudera in Standalone Mode, (II) In CDP Public Cloud [1] & CDP Private Cloud [2], CDE (Cloudera Data Engineering) uses Airflow. Any issues encountered with CDE Airflow on CDP Public Cloud & CDP Private Cloud is Supported. (III) CDE allows External Airflow to be used as well. Yet, Supportability is restricted to Cloudera Airflow Providers only. Refer [3]. If your queries are addressed, Feel free to mark the Post as Solved. Regards, Smarak [1] https://docs.cloudera.com/data-engineering/cloud/orchestrate-workflows/topics/cde-airflow-dag-pipeline.html [2] https://docs.cloudera.com/data-engineering/1.3.4/orchestrate-workflows/topics/cde-airflow-dag-pipeline.html [3] https://docs.cloudera.com/data-engineering/cloud/orchestrate-workflows/topics/cde-airflow-provider.html?
... View more
04-14-2022
01:12 AM
Hello @yagoaparecidoti Thanks for using Cloudera Community. Based on the Post, You encountered "Master Is Initializing" & ended up fixing the same using Command found on Internet. You ended up with few Tables in Broken State, requiring you to delete & recreate them. For documenting the Post, the Issue observed is tracked via [1]. In the HMaster Logs, We should see HBase:Meta & HBase:Namespace Region, which aren't being assigned. Sample tracing shared via [1]. Once the same happens, Using HBCK2 Jar to assign the Region being reported as Unassigned is required. The Command to be used via HBCK2 Jar to assign the Region is shared via [1] as well. The HBCK2 Command has other Options, which (If performed without any oversight) may cause issues with HBase Table availability. The "Master Is Initializing" Error Fix as shared via [1] shouldn't cause any Table to be in Broken State & without knowing the explicit details into the Broken State, It's harder to confirm if Delete-Recreate was the ONLY Way Or, We could have restored the Table in other ways. Having said that, the Post covers "Master Is Initializing" Error & the same has been addressed. As such, We shall mark the Post as Closed. Feel free to share your observation with the Tables in Broken State in a New Post, if you wish to engage for the Community observation & feedback. Regards, Smarak [1] https://github.com/apache/hbase-operator-tools/tree/master/hbase-hbck2#master-startup-cannot-progress-in-holding-pattern-until-region-onlined
... View more
04-14-2022
12:58 AM
Greetings @yagoaparecidoti Thanks for using Cloudera Community. Based on the Post, you wish to confirm if the "maxClientCnxns" Parameter set via CDM (Assuming you mean CM) for ZooKeeper Service is referred in any file associated. You may refer to the Value of "maxClientCnxns" in "zoo.cfg" file associated with the ZooKeeper Process. In CDP/CDH (Managed by CM), You may refer to the same within the ZooKeeper Process Directory under "/var/run/cloudera-scm-agent/process/<ZooKeeper-Process-Directory>/zoo.cfg". The concerned file would be referred in the "ps" Output of the ZooKeeper Process as well. Regards, Smarak
... View more
04-08-2022
01:06 AM
Greetings @wazzu62 We wish to check if you have reviewed @araujo ask for further checks on the concerned issue. If required, Change the Port for the ATS HBase from 17020 to any Value to see if the same helps, assuming the Port is configured to accept request. Regards, Smarak
... View more
04-08-2022
01:00 AM
Hello @Girija Thanks for using Cloudera Community. Since the Post is an Older one, Wish to confirm if you have resolved the Issue. If Yes, Kindly assist by sharing the Solution for wider community audience. Wish to check if your Team had the Kerberos Ticket before submitting the request. Regards, Smarak
... View more
04-08-2022
12:51 AM
Greetings @stephen_obrien Thanks for using Cloudera Community. We see your Team is working with our Support Team for the concerned issue. Based on the Support engagement, We shall update the Post accordingly. Regards, Smarak
... View more
04-08-2022
12:46 AM
Hello @MadhuNP Thanks for using Cloudera Community. We see your Team is working with our Support Team for the concerned issue. Based on the Support engagement, We shall update the Post accordingly. Regards, Smarak
... View more
04-08-2022
12:33 AM
Hello @Neil_1992 & @maykiwogno While we wait for our Nifi Guru @MattWho review, Wish to provide a bit of information on the Lucene Exception. It appears Nifi Provenance Repository uses Lucene for indexing & the AlreadyClosedException means the Lucene Core being accessed has been Closed already, owing to FileSystemException with "Too Many Open Files" for the one of the Core Content "/provenance_repo/provenance_repository/lucene-8-index-1647749380623/_vd_Lucene80_0.dvd". Once AlreadyClosedException is reported, Restarting the Lucene Service would ensure the Cores are initialized afresh. Wish to check if your Team have attempted to increase the OpenFileLimit of the User running the Nifi Process to manage the FileSystemException with "Too Many Open Files" & restart Nifi, which I assume would restart the Lucene Cores as well. Note that the above answer is provided from Lucene perspective as I am not a Nifi Expert. My only intention to get your team unblocked, if the issue is preventing any Nifi concerns. Regards, Smarak
... View more
04-08-2022
12:22 AM
Hello @ISC Thanks for using Cloudera Community. Based on the Post, You are experiencing an Error as shared while using Python with Spark. We shall need full trace of the Error along with which Operation cause the same (Even though the Operation is apparent in the trace shared). Along with the full trace, the Client used (Example: pySpark) & the CDP/CDH/HDP release used. The above details would help us review your Issue & proceed accordingly. Regards, Smarak
... View more
04-08-2022
12:17 AM
Hello @AzfarB We hope the above Post has helped answer your concerns & offered an Action Plan to further review. We are marking the Post as Resolved for now. For any concerns, Feel free to post your ask in a Post & we shall get back to you accordingly. Regards, Smarak
... View more