Member since
02-01-2022
274
Posts
97
Kudos Received
60
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
438 | 05-15-2025 05:45 AM | |
3467 | 06-12-2024 06:43 AM | |
6086 | 04-12-2024 06:05 AM | |
4174 | 12-07-2023 04:50 AM | |
2237 | 12-05-2023 06:22 AM |
08-18-2023
05:35 AM
@sahil0915 It is not clear what you are asking for here. Using Nifi to do this replication, you would be well aware of any records that fail as that is the nature of nifi and how nifi works. NiFi data flows capture failures so you could easily be aware of any records that did not make if from dc1 to dc2 and dc3. Additionally, nifi handles retries, so a replication flow should be resilient to failures, and notify you at that time versus having to fully audit it after replication. If you are using a database technology that replicates across regions or some other replication method and intend to use nifi to check if the replication is complete or accurate, you are going to need to make a nifi flow that will pull all 3 data sets and compare. At the 100 million record row this could be a pretty heavy process with 3 copies of all data coming into NiFi. It would make more sense to me to allow nifi to handle the replication as described above and take the inherit fault tolerance.
... View more
08-18-2023
05:27 AM
A solution and a jira, excellent work gents!
... View more
08-17-2023
06:07 AM
@sree21 You should be able to download the driver and use it anywhere. The license is just for hive endpoint itself. You can find that download page here: https://www.cloudera.com/downloads/connectors/hive/odbc/2-6-1.html
... View more
08-16-2023
06:20 AM
1 Kudo
@janvit04 The pattern you need is this: ${input_date:toDate("MM/dd/yyyy hh:mm:ss"):format("yyyy-MM-dd HH:mm:ss")} I did this in a test which you can find here. In this example I have UpdateAttribute with an input attribute called input_date and its string value is "8/6/2023 12:46 am". In next UpdateAttribute i do the toDate and format. With this setup you may need to modify the format in toDate function to match your input string until it gets right format. For example I thought it should be m/d/yyyy but i got right output using MM/dd/yyyy.
... View more
08-15-2023
06:45 AM
@learner-loading I did some internal digging on this topic and came up with this too: https://docs.cloudera.com/cfm/2.1.3/site-to-site/cdf-datahub-site-to-site.pdf This is more modern doc on using CFM to accomplish site to site with NIFI.
... View more
08-15-2023
05:13 AM
1 Kudo
@learner-loading The Cloudera options here are: CDP Public Cloud (NiFI on Flow Management Data Hub - VMs) - (docs) CDP Public Cloud DataFlow (NiFi on kubernetes) - (docs) CDP Private Cloud Base There are cost calculators in each product but these are dependent on sizing and usage so its not a simple answer. You can open a sales request or reach out to me directly if you need more details. There is a great deal of value add in our offerings in the NiFi space. Including but not limited to: enterprise support and security, cloudera only processors, deploy on-prem, on azure, aws, or gcp. Although these are not SAAS offerings, the Dataflow service should be investigated. The capabilities within this are above and beyond what is found in nifi. Those include, cloud flow catalog, flow designer ui, ready flows, data flow functions (AWS Lamba, Google Cloud Functions, Azure Functions), and more. I am happy to demo and discuss these things as well. Official Sales Requests can be submitted here: https://www.cloudera.com/contact-sales.html
... View more
08-15-2023
04:54 AM
A quick dig shows this port 7191 is for parcel distribution and internal only Peer-to-peer parcel distribution 7190, 7191 Hosts > All Hosts > Configuration > P2P Parcel Distribution Port Used to distribute parcels to cluster hosts during installation and upgrade operations. Reference https://docs.cloudera.com/cdp-private-cloud-base/7.1.6/installation/topics/cdpdc-ports-used-by-cm.html
... View more
08-14-2023
05:54 AM
@Kiranq No, you do not have to use the USE statement. Some things to consider: Make sure each DBCP is configured correctly with the right schema or namespace if possible in the confIguration or connection string. You can also make sure the schema or namespace above your table is included in the processor configuration. For example in PutDatabaseRecord you will find Schema Name in the properties. It could also be done in the sql statement without a specific use statement like select * from default.tablename; assuming your user has access to this default namespace with the tablname.
... View more
08-14-2023
05:43 AM
@StatistiX If you pay attention to the past two CDP Releases (7.1.7 august 2021, 7.1.8 august 2022) you can get an idea of timing for 7.1.9. That said 7.1.9 is expected very soon and also expected include support for redhat 8.8. Due to not knowing what if anything is impacted in 7.1.7 or 7.1.8, i would recommend waiting for official release of 7.1.9 for rhel 8.8 upgrade. If you need official response on using redhat 8.8 with current versions, please open an Upgrade Planning Request ticket with Cloudera. This process will include evaluation of current environment and allow you to have that conversation within your support rep.
... View more
08-08-2023
09:27 AM
@kuhbrille When i execute this in my ranger and check the call, the date is URL parsed: 08%2F08%2F2023 https://host.host.a465-9q4k.cloudera.site/se-aw-mdl/cdp-proxy/ranger/service/assets/accessAudit?page=0&pageSize=25&total_pages=57997&totalCount=1449906&startIndex=0&excludeServiceUser=false&startDate=08%2F08%2F2023&accessResult=0&_=1691511837287 Does that help get your call to take it the date?
... View more