Member since
06-26-2015
505
Posts
129
Kudos Received
114
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
841 | 09-20-2022 03:33 PM | |
2660 | 09-19-2022 04:47 PM | |
1524 | 09-11-2022 05:01 PM | |
1645 | 09-06-2022 02:23 PM | |
2490 | 09-06-2022 04:30 AM |
03-22-2024
01:48 AM
@nanda_bigdata you can install a newer version of impacket. Also, you will have to installed Rust compiler.
... View more
03-11-2024
05:56 AM
1 Kudo
@johnnyyqzheng Apologies for the delayed update. Writing this update as soon as I could. From your previous post, I see that you are still observing the presence of the log4j files after upgrading to CDP 717 SP1. May I know if you got a chance to review the KB article regarding this? If not, Please read this and let me know if you have followup questions. https://my.cloudera.com/knowledge/Clarification-of-Log4J-1x-Remediation-on-CDP-717-SP1-and-CM771?id=373883 V
... View more
03-06-2024
12:38 AM
It seems like there might be an issue with the way you're using single quotes in the loop. The variable eachline should be expanded, but it won't if it's enclosed in single quotes. Try using double quotes around the variable and see if that resolves the issue. Here's the corrected loop: for eachline in "${testarray[@]}"
do
beeline -f "${eachline}.sql"
done This way, the value of eachline will be correctly expanded when constructing the command. Also, ensure that the SQL files are present in the correct path or provide the full path if needed. If the issue persists, please provide more details on the error message or behavior you're experiencing for further assistance.
... View more
02-07-2024
11:34 AM
1 Kudo
@SS-dev Welcome to the Cloudera Community!
As this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post. Thanks.
... View more
10-08-2023
09:43 AM
I think you faced with NiFi Cross-Site Request Forgery Protection. Here is the link to manual: https://nifi.apache.org/docs/nifi-docs/html/administration-guide.html#csrf-protection When you execute authorization request (/access/token) you got two cookies: __Secure-Authorization-Bearer and __Secure-Request-Token. Value of __Secure-Authorization-Bearer cookie you can use as value for Authorization header or maybe your http client can add it automatically. And to avoid CSRF security filter you should add to all POST/PUT/DELETE requests header with name Request-Token and value of __Secure-Request-Token cookie.
... View more
10-03-2023
08:12 AM
@Kiranq, What I found is that the JoltTransformRecord expects only single record to work with hence the name. I noticed when I try to pass an array I was getting the error "...error transforming the first record", however if I pass just one json record it works. If you have an array of json\csv and you are looking to split and process each record individually then I would suggest that you split the records before the JoltTransformRecord. If you dont want to split the array then I recommend using JoltTranformJson first and then use Convert Record processor to convert to CSV.
... View more
09-13-2023
07:00 AM
Hello @hebamahmoud If the issue is has been from any of the above responses, Could you accept it as a solution? Regards, Chethan YM
... View more
09-12-2023
11:14 PM
1 Kudo
We are pleased to announce the general availability of Cloudera Streaming Analytics (CSA) 1.11 on CDP Private Cloud Base 7.1.9. This release includes improvements to SQL Stream Builder (SSB) as well as updates to Flink 1.16.2. These changes are focused on enhancing the user experience and fixing bugs, making the product more robust and stable. Sincere thanks to all the individuals who helped with this release and did an incredible job to get this ready. Key features for this release Rebase to Apache Flink 1.16.2 - Apache Flink 1.16.2 is now supported in CSA 1.11. Apache Iceberg support - Support for Apache Iceberg tables using Iceberg v2 format has been added to Flink and SSB. For more information, see the Creating Iceberg tables documentation. Links What's New in CSA 1.11 Documentation Iceberg Tables Iceberg Connector REST API v2 Reference BLOG: Building a Stateful Intrusion Detection System with SSB
... View more
Labels:
08-30-2023
01:08 AM
Anyone else experienced this issue ? is there a solution that does not involve creating one CaptureChangeMySQL processor instance per source table ?
... View more