Member since
02-01-2022
274
Posts
97
Kudos Received
60
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
440 | 05-15-2025 05:45 AM | |
3481 | 06-12-2024 06:43 AM | |
6124 | 04-12-2024 06:05 AM | |
4180 | 12-07-2023 04:50 AM | |
2247 | 12-05-2023 06:22 AM |
05-12-2023
08:40 AM
Can you please show the source json in a : code box Also, try a manual test, without attribute, such as: SELECT *
FROM FLOWFILE
WHERE RPATH(data, '/room') = 'A' This will ensure the query is correct to the json payload. Once that works, start testing adding the attribute.
... View more
05-12-2023
06:59 AM
@rafy I would try this: SELECT * FROM FLOWFILE WHERE RPATH_STRING(data, '/room')='${ip}' Assuming ip is an attribute (${ip}) ofcourse.
... View more
05-11-2023
05:42 AM
1 Kudo
@ushasri I believe the solution here is to use the Record based processors with a specified schema. This allows you to provide the correct schema to the Reader and the Writer regardless of the field names in original excel data source.
... View more
05-10-2023
05:56 AM
@zzeng Great article. Reach out to me on internal channels. I would love to show you my oracle to kudu demo, using kafka and schema registry.
... View more
05-10-2023
05:42 AM
1 Kudo
@SandyClouds ^^ that is how you do it. One important thing to mention. That processor is not meant to be deployed in a set it up and forget it situation. CDC against another database requires a lot more technical attention to how the data is changing over time AFTER you grab data for first time. In my opinion, this processor is meant to be a conversation starter, or a way to take a 1 time shot of data source, where you may be watching it run, but not expecting it to run indefinitely and keep 2 systems synced.
... View more
05-10-2023
05:36 AM
@ryu Excellent questions here. I will address each below: So when I am asked like in an interview or something, they ask what CDH version do you use, and when i say CDH 7.2.16, the people interviewing me asks "are you sure?" Yes, you are sure. Even new versions of CDP, have CDH in the artifact filenames: 7.1.4-1.cdh7.1.4.p37.14288300 So is there a CDP version of CDH vs maybe on prem version of CDH Current versions of our platform are CDP 7.x. Previous versions are CDH 6.x and older. There is On Prem CDP called CDP Private Cloud Base. There is Public Cloud CDP called CDP Public Cloud in Aws, Azure, and GCP. Is there some difference in versioning between on prem or CDP etc? This is my favorite question. The differences between CDP on Prem and CDP in the cloud are going away quickly. It is part of our modern data architecture for workload movement from on prem to the cloud to have the least amount of differences as possible. As such, these 2 different form factors of CDP are getting closer and closer to parity.
... View more
05-10-2023
05:24 AM
@mwblee HDP is no longer supported platform. You cannot access the final HDP artifacts without cloudera subscription. I would highly recommend that you take a look at CDP and modern supported versions of original HDP components.
... View more
05-08-2023
07:24 AM
@sridharavulapat You should be able to find the required hive values in hive-site.xml file. You can get this file from Cloudera Manager. Additionally, use Cloudera Manager to download hive drivers and get the fully qualified jdbc url.
... View more
05-04-2023
05:56 AM
@anony I believe your solution is to put/keep Response Data as flowfile content. Then you can do QueryRecord or PartitionRecord to iterate through each "name" object in the Response Data array. Then downstream you are able to use EvalateJsonPath to get the json object values into attributes or flowfile content. For example name= $.name, nameID = $.Speciality[0].nameId, etc
... View more
05-04-2023
05:47 AM
1 Kudo
@ushasri You should check out NiFI Registry: https://nifi.apache.org/registry.html Using NiFi Registry you are able to version control flows during development. Using same Nifi Registry you can now deploy these flows to other environment(s) such as your Cloudera licensed version (HDF,CDF,CFM). Additionally, XML templates are going away. In modern versions of nifi, you should use the Create Flow Definition, and transfer these JSON definition files manually between environments and/or nifi developers. Other CI/CD DFLC (Data Flow Lifecycle) concepts for deploying flows across environments are using NIFI CLI API programmatically to do manual deployment operations. We also often seen some level of integration between Github and NiFi Registry.
... View more