Member since
02-01-2022
281
Posts
103
Kudos Received
60
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1120 | 05-15-2025 05:45 AM | |
| 4949 | 06-12-2024 06:43 AM | |
| 7921 | 04-12-2024 06:05 AM | |
| 5830 | 12-07-2023 04:50 AM | |
| 3203 | 12-05-2023 06:22 AM |
03-21-2023
06:16 AM
Yes, I am doing the same. And how to maintain the order of the dynamic properties in Invokehttp processor as the order is also important else it is saying bad request. Whenever i am adding those, these are automatically arranging the alphabetical orders instead the order on which i am adding
... View more
03-20-2023
05:48 AM
@Fahmihamzah84 This appears to be an issue with your schema. The BigQuery error is suggesting an issue trying to cast a string into a collection (array/list/ect). It's hard to tell which array may be causing the issue as there are many. My suggestion is to set the processor to log level DEBUG and see if you can get more verbose error. This will help you figure out which field or fields is the culprit. Keep in mind it could be one of the empty arrays too. I do not suggest the following as a solution just as path to figuring out where the problem is. Sometimes when i have issues with type casting, i make everything a string temporarily and for development. If you do this carefully one at a time, when the error goes away, you can determine which field it is. This also helps you identify a working state for your flow and allow you to work from that operational base to find solution for the end schema being the format you need.
... View more
03-16-2023
08:16 AM
1 Kudo
Awesome news, +2 solutions here.
... View more
03-02-2023
06:44 AM
1 Kudo
@fahed What you see with the CDP Public Cloud Data Hubs using GCS (or object store) is a modernization of the platform around object storage. This removes differences across aws, azure, and on-prem (when Ozone is used). It is a change by customer demand so that workloads are able to be built and deployed with minimal changes from on prem to cloud or cloud to cloud. Unfortunately that creates a difference you describe above, but those are risks we are willing to take ourselves in favor of modern data architecture. If you are looking for performance, you should take a look at some of the newer options for databases: impala and kudu (this one uses local disk). Also we have Iceberg coming into this space too.
... View more
03-01-2023
06:33 AM
1 Kudo
I increased the size of the ListenUDP processor and turned to false the content.repository.archive property and everything works again. Thanks for your reply
... View more
03-01-2023
04:14 AM
Nice and Quick! Excellent!
... View more
02-24-2023
05:55 AM
1 Kudo
@kishan1 In order to restart a specific processor group you will need to use some command line magic against the Nifi API. For example, this could be done by using a command to stop the processor group, then the restart nifi command, then start processor group. You can certainly be creative in how you handle that approache once you have experimented with the API. https://nifi.apache.org/docs/nifi-docs/rest-api/index.html
... View more
02-23-2023
05:09 AM
1 Kudo
@fahed That size is to be able to grow and serve in production manner. At first that disk usage could be low. For DataHubs, My recommendation is to start small and grow as needed. Most of your work load data should be in object store(s) for the data hubs, so dont think of that "hdfs" disk as being size constrained to initial creations of the hub.
... View more
02-22-2023
08:28 AM
1 Kudo
@merlioncurry Lacking a bit of deatils, so making some assumptions that you used an Ambari UI to upload to HDFS. So those files are going to be in hdfs://users/maria_dev, not on the actual machine location for the same users. You will need use hdfs commands to view them. If they do not work, then the path you uploaded may be different. From the sandbox prompt: hdfs dfs -ls /users/ hdfs dfs -ls /users/maria_dev
... View more