Member since
11-17-2021
1149
Posts
258
Kudos Received
30
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 171 | 04-23-2026 02:02 PM | |
| 555 | 03-17-2026 05:26 PM | |
| 5232 | 11-05-2025 10:13 AM | |
| 875 | 10-16-2025 02:45 PM | |
| 1459 | 10-06-2025 01:01 PM |
08-06-2022
09:02 AM
@shrikantbm& team, Yes, in this case, we need to check cleanup.policy of the topic __consumer_offsets. If the existing cleanup.policy=compact then the log segment of this topic will not be deleted. You should follow the below steps to conclude and resolve this issue initially. 1) Check what is current cleanup.policy of the topic __consumer_offsets. You can check it using the command: kafka-topics.sh --bootstrap-server <broker-hostname:9092> --describe OR kafka-topics.sh --zookeeper <zookeeper-hostname:2181> --describe --topics-with-overrides Note: topic_name is the name for which you are facing an issue 2) If you want to clear the old log segment of this topic, then you should set cleanup.policy like cleanup.policy=compact,delete,retention.ms=<30days> compact = when the kafka-log is rolled over, it will be compacted delete - once the offset.retention.ms is reached, the older logs will be removed retention.ms=<30days> > the old log segment will be deleted after 30 days. Note: 30 days are just an example here and this setting will be in ms. You should set it as per your requirement after checking it with the application team and their need. For "delete", the property "log.cleaner.enable" must be set to "true" After configuring this cleanup policy data will be deleted as per retention.ms as suggested above. If you will not set retention.ms then old log segment will be deleted as per retention period set in the CM / Ambari >> kafka >> Conf. The setting is log.retention.hours = <7 Days default> in CM >> Kafka, check what it is in your case so that log segment older than 7 days will be deleted. Kafka will keep checking the old log segment with the help of the property log.retention.check.interval.ms . Important note: The "delete" on consumer offsets is that you may lose offsets which can lead to duplication/data loss. So check it with your application team before setting a deletion policy. 3) If you still face the same issue, then broker logs need to be reviewed for the root cause of the issue and make the changes accordingly. If you found this information helped with your query, please take a moment to log in and click on KUDOS 🙂 and "Accept as Solution" below this post. Thank you.
... View more
07-26-2022
07:06 AM
@jeyaguna Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks
... View more
07-26-2022
06:47 AM
@Ploeplse Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks
... View more
07-20-2022
09:56 AM
Here are some highlights from the month of June
206 new support questions
3 new community articles
646 new members
Rank
Community Article
Author
Components/ Labels
#1
NiFi cluster sandbox on Docker
@araujo
Apache NiFi
#2
Decompressing nested ZIP files in NiFi
@araujo
Apache NiFi
#3
Cloudera Machine Learning adds ability for admins to add a static subdomain for the CML Workspace
@cpv0310
Cloudera Machine Learning (CML)
#4
MySQL CDC with Kafka Connect/Debezium in CDP Public Cloud
@cnelson2
Apache Kafka
Cloudera Data Platform (CDP)
Kerberos
Thank you to @shehbazk for supporting our community! :trophy: 👏 Check out the Community Member Spotlight on Cloudera Linkedin!
We would like to recognize the below community members and employees for their efforts over the last month to provide community solutions.
See all our top participants at Top Solution Authors leaderboard and all the other leaderboards on our Leaderboards and Badges page.
@SAMSAL @rafy @snm1523 @Luwi @araujo @MattWho @jagadeesan @rki_
Share your expertise and answer some of the below open questions. Also, be sure to bookmark the unanswered question page to find additional open questions.
Unanswered Community Post
Components/ Labels
Central management of connection strings - Sqoop - HDP
Apache Ambari Apache Hadoop Apache Sqoop
Spark Streaming job not reading data from Kafka
Apache Kafka
Apache Spark
Cloudera Data Platform (CDP)
How do I drop empty Impala partitions?
Apache Hive
Apache Impala
Cloudera Enterprise Data Hub
... View more
07-20-2022
08:46 AM
1 Kudo
@AnilKumar7 Hi, I have reached to you via PM for further details.
... View more
07-14-2022
11:55 PM
@data_diver Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks
... View more
07-13-2022
11:55 AM
I asked another colleague to open the non-technical ticket. Let's see if it resolves. Ideally, this adjustment would be requested by myself, as I own the company's @dominio account.
... View more
07-12-2022
03:08 PM
@Chakkara As far as i remember , distributed cache is not having consistency . You could use Hbase or HDFS for storing the status of success or failure of the processors for downstream application. Once you saved the Success and Failure at Hbase . You can retrieve it from the Hbase processor using the row ID. Build a REST API NiFi flow to pull the status from Hbase for example HandleHTTPRequest --> FetchHbaseRow - HandleHTTPResponse You can call the HTTP API (Request and Response) via shell script/curl and call the script from Control-M.
... View more
07-11-2022
11:52 AM
@yagoaparecidoti Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks
... View more
07-07-2022
12:53 PM
@Meshal Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks
... View more