Member since
01-16-2018
418
Posts
24
Kudos Received
50
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
117 | 06-22-2022 12:15 AM | |
144 | 06-21-2022 11:38 PM | |
151 | 06-21-2022 11:17 PM | |
238 | 04-14-2022 01:29 AM | |
818 | 04-14-2022 01:12 AM |
08-08-2022
02:45 AM
Hello @hbasetest You wish to enable Normalizer at Cluster Level irrespective of the Table Level Setting i.e. NORMALIZATION_ENABLED be True or False. As far as I believe, We would require Table Level enabling. Having said that, If you can Open a Post on the same by using the Steps shared by @VidyaSargur, Our fellow Community Gurus can get back to you sooner, as compared to a Comment on an Article written in 2016.
... View more
08-04-2022
02:28 AM
Hello @achandra, This is an Old Post yet closing the same by sharing the feedback concerning your ask for wider audience. The API is failing owing to Space between "NOW-" & "7DAYS". There shouldn't be any gap between the same. In Summary, the Command is below, where Customer needs to set the HTTP(s) header, Solr Host & Solr Port accordingly. Additionally, the Example uses "ranger_audits" Collection & "evtTime" field to delete any Documents older than 7 Days: ### curl -k --negotiate -u : "http[s]://<Any Solr Host FQDN>:<Solr Port>/solr/ranger_audits/update?commit=true" -H "Content-Type: text/xml" --data-binary "<delete><query>evtTime:[* TO NOW-7DAYS]</query></delete>" Regards, Smarak
... View more
07-06-2022
07:49 AM
Hello @LakshmiSegu We hope your Q concerning the Zeppelin Access issue is addressed by our 06/21 Post. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
07-06-2022
07:47 AM
1 Kudo
Hi @cprakash Kindly review our Post on 06/21 & share your observation. If you have resolved the issue, Kindly share the Solution for our wider Community Members. Regards, Smarak
... View more
07-06-2022
07:41 AM
Hello @sgenchev We hope your Q concerning identifying the Region/RegionServers was addressed by our 06/22 Post. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
07-06-2022
07:40 AM
Hello @caisch We hope our Post on 06/21 help answer your query concerning Solr TTL. As such, We shall mark the Post as Resolved. If you have any concerns, Feel free to update the Post & we shall get back to you accordingly. Regards, Smarak
... View more
06-22-2022
12:15 AM
Hello @sgenchev While I am not familiar with building Charts to answer the Query, I shall share a few CLI approaches: (I) HBCK: Running "hbase hbck --details" prints [1] detailing the Regions per RegionServer. (II) Using JMX: Say, the HMaster UI is "http://HMaster1:22002/master-status". Replace "master-status" with "jmx", which prints the JSON Output. Additional filtering can be done to capture the Regions per RegionServers. Regards, Smarak [1] HBCK Output: RegionServer: RegionServer1,22101,1655811831192 number of regions: 1 RegionServer: RegionServer2,22101,1655811830393 number of regions: 1 RegionServer: RegionServer2,22101,1655811832307 number of regions: 1
... View more
06-21-2022
11:43 PM
Hello @cprakash Let us know if you have fixed the ConnectionRefused Error to Master1 on Port 16000. As @rki_ mentioned, Review the HMaster Logs to confirm the reasoning for ConnectionRefused. Few possible scenarios being Port 16000 is being used by any other Service Or, "master1" isn't correctly being mapped as per DNS Or, Port 16000 may be blocked. Based on the HMaster Logs, We may offer further review. As such, If the Issue persists, Share the HMaster Logs covering the concerned Stacktrace as well. If the Issue has been resolved, We shall appreciate if you can share the details into the Cause & Fix applied. Regards, Smarak
... View more
06-21-2022
11:38 PM
Hello @LakshmiSegu We hope your query was addressed by Shehbaz's response. In Summary, (I) Ensure you Username has an IDBroker Mapping (Actions > Manage Access > IDBroker Mappings). (II) Include the "spark.yarn.access.hadoopFileSystems" Parameter to point to the S3 Path [1]. Regards, Smarak [1] https://docs.cloudera.com/runtime/7.2.15/developing-spark-applications/topics/spark-s3.html
... View more
06-21-2022
11:17 PM
Hello @caisch Thanks for using Cloudera Community. Based on your Post, you wish to confirm if TTL for HDFS can be set to 90Days & other Services' TTL is set to 14Days. Since you selected Solr, Let me answer your Post & you can let me know if I understood the Post differently. In Solr, We have Collections which may have Data on HDFS or Local. For Example, RangerAudits Collection may have Data on HDFS & Atlas Collection may have Data on Local. At each Collection Level, the "solrconfig.xml" captures the TTL [1] via DocExpirationUpdateProcessorFactory Class. You can configure the TTL at each Collection Level in Solr & they would Cleanup the Underlying Data, be it on HDFS or Local. Using the above Example of RangerAudits using HDFS Storage & Atlas using Local Storage, We can set RangerAudits to expire at 90 Days & Atlas to expire at 14 Days, which in turn would remove the Underlying Data from HDFS & Local respectively for RangerAudits & Atlas. Kindly review & let us know if the above answers your Post. If No, You may clarify & we shall get back to you accordingly. Regards, Smarak [1] https://solr.apache.org/docs/8_4_0/solr-core/org/apache/solr/update/processor/DocExpirationUpdateProcessorFactory.html
... View more
04-19-2022
01:07 AM
Hello @SVK If your queries concerning Apache Airflow has been addressed, Feel free to mark the Post as Solved. If you have any further ask, Kindly share the same & we shall get back to you accordingly. Regards, Smarak
... View more
04-19-2022
01:05 AM
Hello @HiThere We hope your query concerning VMs & recommendations around Hardware is answered. We are marking the Post as Closed. If you have any further concerns, Feel free to post your ask & we shall answer your queries. Regards, Smarak
... View more
04-14-2022
01:29 AM
Hello @HiThere Thanks for using Cloudera Community. To your Query, Kindly refer [1] & [2] for Resource Requirement for CDP v7.1.7. Note that Documentation refers to Hardware requirement in terms of Resources (CPU, Memory, Network, Disk) as opposed to Physical Machines or Virtual Machines. As long as your Team meet the Hardware requirement for Storage & Compute, the Virtualized & Bare-Metal choices shouldn't matter. Regards, Smarak [1] https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/installation/topics/cdpdc-hardware-requirements.html [2] https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/concepts/topics/cm-vpc-networking.html
... View more
04-14-2022
01:23 AM
Hello @SVK Thanks for using Cloudera Community. Based on your post, You wish to confirm if Airflow is Supported by Cloudera. Summarizing the responses shared by my Colleagues for you: (I) Airflow isn't Supported by Cloudera in Standalone Mode, (II) In CDP Public Cloud [1] & CDP Private Cloud [2], CDE (Cloudera Data Engineering) uses Airflow. Any issues encountered with CDE Airflow on CDP Public Cloud & CDP Private Cloud is Supported. (III) CDE allows External Airflow to be used as well. Yet, Supportability is restricted to Cloudera Airflow Providers only. Refer [3]. If your queries are addressed, Feel free to mark the Post as Solved. Regards, Smarak [1] https://docs.cloudera.com/data-engineering/cloud/orchestrate-workflows/topics/cde-airflow-dag-pipeline.html [2] https://docs.cloudera.com/data-engineering/1.3.4/orchestrate-workflows/topics/cde-airflow-dag-pipeline.html [3] https://docs.cloudera.com/data-engineering/cloud/orchestrate-workflows/topics/cde-airflow-provider.html?
... View more
04-14-2022
01:12 AM
Hello @yagoaparecidoti Thanks for using Cloudera Community. Based on the Post, You encountered "Master Is Initializing" & ended up fixing the same using Command found on Internet. You ended up with few Tables in Broken State, requiring you to delete & recreate them. For documenting the Post, the Issue observed is tracked via [1]. In the HMaster Logs, We should see HBase:Meta & HBase:Namespace Region, which aren't being assigned. Sample tracing shared via [1]. Once the same happens, Using HBCK2 Jar to assign the Region being reported as Unassigned is required. The Command to be used via HBCK2 Jar to assign the Region is shared via [1] as well. The HBCK2 Command has other Options, which (If performed without any oversight) may cause issues with HBase Table availability. The "Master Is Initializing" Error Fix as shared via [1] shouldn't cause any Table to be in Broken State & without knowing the explicit details into the Broken State, It's harder to confirm if Delete-Recreate was the ONLY Way Or, We could have restored the Table in other ways. Having said that, the Post covers "Master Is Initializing" Error & the same has been addressed. As such, We shall mark the Post as Closed. Feel free to share your observation with the Tables in Broken State in a New Post, if you wish to engage for the Community observation & feedback. Regards, Smarak [1] https://github.com/apache/hbase-operator-tools/tree/master/hbase-hbck2#master-startup-cannot-progress-in-holding-pattern-until-region-onlined
... View more
04-14-2022
12:58 AM
Greetings @yagoaparecidoti Thanks for using Cloudera Community. Based on the Post, you wish to confirm if the "maxClientCnxns" Parameter set via CDM (Assuming you mean CM) for ZooKeeper Service is referred in any file associated. You may refer to the Value of "maxClientCnxns" in "zoo.cfg" file associated with the ZooKeeper Process. In CDP/CDH (Managed by CM), You may refer to the same within the ZooKeeper Process Directory under "/var/run/cloudera-scm-agent/process/<ZooKeeper-Process-Directory>/zoo.cfg". The concerned file would be referred in the "ps" Output of the ZooKeeper Process as well. Regards, Smarak
... View more
04-08-2022
01:06 AM
Greetings @wazzu62 We wish to check if you have reviewed @araujo ask for further checks on the concerned issue. If required, Change the Port for the ATS HBase from 17020 to any Value to see if the same helps, assuming the Port is configured to accept request. Regards, Smarak
... View more
04-08-2022
01:00 AM
Hello @Girija Thanks for using Cloudera Community. Since the Post is an Older one, Wish to confirm if you have resolved the Issue. If Yes, Kindly assist by sharing the Solution for wider community audience. Wish to check if your Team had the Kerberos Ticket before submitting the request. Regards, Smarak
... View more
04-08-2022
12:51 AM
Greetings @stephen_obrien Thanks for using Cloudera Community. We see your Team is working with our Support Team for the concerned issue. Based on the Support engagement, We shall update the Post accordingly. Regards, Smarak
... View more
04-08-2022
12:46 AM
Hello @MadhuNP Thanks for using Cloudera Community. We see your Team is working with our Support Team for the concerned issue. Based on the Support engagement, We shall update the Post accordingly. Regards, Smarak
... View more
04-08-2022
12:33 AM
Hello @Neil_1992 & @maykiwogno While we wait for our Nifi Guru @MattWho review, Wish to provide a bit of information on the Lucene Exception. It appears Nifi Provenance Repository uses Lucene for indexing & the AlreadyClosedException means the Lucene Core being accessed has been Closed already, owing to FileSystemException with "Too Many Open Files" for the one of the Core Content "/provenance_repo/provenance_repository/lucene-8-index-1647749380623/_vd_Lucene80_0.dvd". Once AlreadyClosedException is reported, Restarting the Lucene Service would ensure the Cores are initialized afresh. Wish to check if your Team have attempted to increase the OpenFileLimit of the User running the Nifi Process to manage the FileSystemException with "Too Many Open Files" & restart Nifi, which I assume would restart the Lucene Cores as well. Note that the above answer is provided from Lucene perspective as I am not a Nifi Expert. My only intention to get your team unblocked, if the issue is preventing any Nifi concerns. Regards, Smarak
... View more
04-08-2022
12:22 AM
Hello @ISC Thanks for using Cloudera Community. Based on the Post, You are experiencing an Error as shared while using Python with Spark. We shall need full trace of the Error along with which Operation cause the same (Even though the Operation is apparent in the trace shared). Along with the full trace, the Client used (Example: pySpark) & the CDP/CDH/HDP release used. The above details would help us review your Issue & proceed accordingly. Regards, Smarak
... View more
04-08-2022
12:17 AM
Hello @AzfarB We hope the above Post has helped answer your concerns & offered an Action Plan to further review. We are marking the Post as Resolved for now. For any concerns, Feel free to post your ask in a Post & we shall get back to you accordingly. Regards, Smarak
... View more
04-06-2022
11:51 PM
Thank You @RangaReddy for this detailed write-up. The level of detailing is awesome 👏
... View more
03-29-2022
01:38 AM
Hello @Suresh_lakavath Since we haven't heard from your side concerning the Post, We are marking the Post as Closed for now. Feel free to Update the Post based on your Team's observation from the Action Plan shared on 03/09. Regards, Smarak
... View more
03-29-2022
01:17 AM
Hello @Moawad Hope you are doing well. Kindly let us know if the Post on 03/20 documenting few Links from CDH v6.x helped your Team. Regards, Smarak
... View more
03-29-2022
01:13 AM
1 Kudo
Hello @dutras As the Issue has been resolved via Support Case, We are marking the Post as resolved. For reference, this Case required Repair & associated Steps, which are verbose to be documented in Community. For anyone facing such issues, Kindly submit a Support Case. Regards, Smarak
... View more
03-29-2022
01:00 AM
Hello @AzfarB Thanks for using Cloudera Community. Based on the Post, Your Team observed Solr-Infra JVM reporting WARNING for Swap Space more than 200MB being utilised. Restarting the Solr-Infra JVM ensured the WARNING went away. Note that Swapping isn't Bad in general & the same has been discussed in detail by Community in [1] & [2]. Plus, Deleting RangerAudits Documents won't affect the same as Solr uses JVM as documented [3]. Indexed Documents aren't persisted in Memory unless Cached, thereby ensuring Deletion won't fix the Swapping guaranteed. As your Screenshot shows, the Host itself is running short on Memory (~99% Utilised) & Overall Swap is ~80% at ~47GB, out of which Solr-Infra is contributing <1GB. As documented in the below Links, Your Team can focus on the Host Level Usage & Considering Increasing the Swap Threshold from 200MB to at least 10% of the Heap i.e. 2GB for a Warning. 01 additional point can be made as to why Solr-Infra Restart helped resolved the WARNING. This needs to be looked at from the Host perspective as to the amount of Memory freed & Whether the Overall Swap Usage reduced at Host Level after Solr-Infra Restart as opposed to Solr-Infra WARNING being suppressed only. Regards, Smarak [1] https://unix.stackexchange.com/questions/2658/why-use-swap-when-there-is-more-than-enough-free-space-in-ram [2] https://chrisdown.name/2018/01/02/in-defence-of-swap.html [3] https://blog.cloudera.com/apache-solr-memory-tuning-for-production/
... View more
03-20-2022
08:51 PM
Hello @Moawad Thanks for using Cloudera Community. Based on the Post, Your Team is having issues connecting Kudu via pySpark. Kindly confirm whether a Simple Example [1] as documented in CDH 6.2.x Guide works for your Team. Regards, Smarak [1] https://docs.cloudera.com/documentation/enterprise/6/6.2/topics/kudu_development.html [2] https://kudu.apache.org/docs/developing.html
... View more
03-17-2022
12:07 AM
Hello @Suresh_lakavath Hope you are doing well. We wish to follow up with you on this Post. Regards, Smarak
... View more