Member since
09-20-2023
13
Posts
3
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
436 | 10-30-2023 07:48 AM |
04-15-2024
06:17 AM
When I run this command via the cdp cli, it just returns a JSON like this: ``` { "archiveName": "[MY_FLOW_NAME].tar.gz" } ``` I want to get the actual flow definition.
... View more
03-26-2024
10:28 AM
I want to use the Atlas API to add/edit user-defined properties to a particular entity. I've previously had success using the API to edit properties that are already defined, such as an entity description: ``` curl --location --request PUT 'https://[host]/api/atlas/v2/entity/guid/[guid]?name=description' \ --header 'Content-Type: application/json' \ --header 'Authorization: Basic' \ --data '"a test table"' ``` However, it's not clear how to use the API to create and edit user-defined properties. When I try changing the url query to a user-defined property name, it complains that that property is not defined for the entity type. For reference, in the UI, this is done fairly easily.
... View more
Labels:
- Labels:
-
Apache Atlas
03-26-2024
07:00 AM
That works. Thanks!
... View more
03-25-2024
05:24 AM
I'd like to have a process for automatically exporting the data flow definition and configuration without needing to go into the CDF UI. Right now, if I wanted the configuration, I'd click here in the UI: My thought was to use the CDP CLI, but under the df section, there doesn't appear to be a command for this. Any suggestions?
... View more
Labels:
03-25-2024
05:15 AM
I'm trying to use the cdp cli (v. 0.9.106 on my local machine) to get KPI metrics from CDF. I'm using the command `cdp df list-deployment-system-metrics --deployment-crn my-deployment-crn` but struggling to correctly format the required `--metrics-time-period` arg. The documentation isn't helpful here: I've tried a few things like: `--metrics-time-period "2024-03-24T12:00:00Z:2024-03-24T16:00:00Z"` `--metrics-time-period LAST_24_HOURS` `--metrics-time-period "yesterday"` But they all scream back: An error occurred: No enum constant com.cloudera.dfx.metrics.TimeSpan.TimePeriod.yesterday (Status Code: 400; Error Code: INVALID_ARGUMENT; Service: df; Operation: listDeploymentSystemMetrics; If someone could elaborate on what this command is expecting, or better yet, offer up some examples of successful executions, that'd be great!
... View more
Labels:
02-29-2024
06:01 AM
Thanks @nhassell , that's resolving the 'cannot locate driver' issue. Now I'm just hitting my head against pesky host/port configuration problems. For the life of me, I cannot get past: ``` [S1000][unixODBC][Cloudera][DriverSupport] (1170) Unexpected response received from server. Please ensure the server host and port specified for the connection are correct. ``` Obviously I'm doing something wrong. But your suggestion solved the issue around the driver library, so thanks.
... View more
02-13-2024
05:03 AM
1 Kudo
No I'm still struggling with ODBC.
... View more
02-13-2024
04:59 AM
1 Kudo
I have an existing iceberg table and I'm trying to evolve schema by adding a new column: `alter table my_db.my_table add column insert_time timestamp` When I execute the above statement, I get a 'New column(s) have been added to the table.' but the column doesn't appear. I tried running a REFRESH, an INVALIDATE METADATA, and I tried restarting Impala but it had no effect. I tried the same operation on a test table and it worked. I'm using Impala as a query engine and Azure for storage. Impala is running in a Data Hub Cluster.
... View more
12-07-2023
12:06 PM
I want to set up a Nifi flow that gets data from a public RSS feed and loads it into a data lake. This RSS feed updates irregularly and when it does update it overwrites previous content. What processor(s) should I use to get data from the RSS feed (close to) when it has updated? Is it as simple as using InvokeHTTP repeatedly, checking for a change in output, then loading into data lake if the content differs from the previous invocation? Is there another way if I don't want to make the HTTP request so frequently?
... View more
Labels:
- Labels:
-
Apache NiFi
-
Cloudera DataFlow (CDF)
10-30-2023
07:48 AM
Resolved. The issue was that I was using an ExecuteSQL processor before the PutIceberg processor and I neglected to use logical types. The fix was to switch the property 'Use Avro Logical Types' to true.
... View more