Member since
07-07-2020
15
Posts
3
Kudos Received
4
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 5215 | 12-14-2020 06:42 AM | |
| 1872 | 11-06-2020 04:57 AM | |
| 13375 | 08-31-2020 05:14 AM | |
| 5036 | 07-20-2020 11:27 PM |
12-14-2020
06:42 AM
1 Kudo
Try to use this link : http://archive.cloudera.com/csa/1.0.0.0/parcels/ I think you need credential and cloudera 7 to use latest versions of CSA. I'm using a trial version on with CM 6.2 and CDH 6.2.1 You can get it from this link https://www.cloudera.com/downloads/cdf/csa-trial.html
... View more
12-13-2020
04:17 AM
Do you have the CSD ? You also need it to manage the service. https://docs.cloudera.com/csa/1.2.0/installation/topics/csa-installing-parcel.html
... View more
11-06-2020
04:57 AM
Hello, I'm not sure to understand your problem, but i'll try to answer to "Files ingested into HDFS but I don't see Files in HDFS". When you process a flowfile with a processor, the processor will route the flowfile to relationship according with what it managed to do with the flowfile. The PutHDFS processor have two relationship : success and failure. If the processor managed to put the flowfile to HDFS, the flowfile is routed to the relationship success and you can continue to process it. This unless you check automatically terminate the relationship success, then NiFi won't do anything more with the flowfile. Same for the failure relationship. Here you automatically terminate both success and failure relationship. For success relationship this is ok, because your file was put to HDFS. But for failure relationship, the file is not put to HDFS and you don't handle the failure. Just route the failure relationship to any other processor, you don't even need to use it at this point, you just want to see if flowfile are route to the failure relationship, and so if the queue corresponding to this relationship is filled. You can also see the failure if a red error box appear in the top right corner of the processor, but previous method if to understand the concept of relationship routes. If so, the flowfile are not put to HDFS, probably because of a bad configuration of the PutHDFS processor. In this case, i'll be happy to help you with this configuration (properties tab).
... View more
08-31-2020
05:14 AM
Hello, You can create a nested node "nifi-content" using the ReplaceText processor. You just need to configure the Replacement Value field : {"nifi-content":$1} Then just add a new node "filename" using the JoltTransformJSON processor with following Jolt specification : [
{
"operation": "default",
"spec": {
"filename": "${filename}"
}
}, {
"operation": "default",
"spec": {
"*": "&"
}
}
] Maybe you can create the "nifi-content" node and add the "filename" node only using one JoltTransform JSON ; but i'm not very experienced using Jolt and I did not find out how.
... View more
07-20-2020
11:27 PM
1 Kudo
You can use Nifi to save your Kafka messages into HDFS (for instance). Something like this : - ConsumeKafka : flowfile content is the Kafka message itself, and you have access to some attributes : topic name, partition, offset, key...(but not timestamp !). When i need it I store the timestamp in the key. - ReplaceText : build your backup line using flowfile content and attributes - MergeContent : to build a big file containing multiple Kafka message - Extracttext : to set attribute to be used as filename - PutHDFS : to save the created file into HDFS And you can do the reverse if you need to push it bash to your kafka cluster.
... View more