Member since
01-02-2020
40
Posts
3
Kudos Received
5
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7302 | 12-23-2020 09:33 AM | |
1922 | 05-18-2020 01:27 AM | |
1716 | 04-28-2020 11:02 AM | |
4803 | 04-23-2020 12:20 PM | |
2310 | 01-25-2020 11:50 PM |
01-13-2023
07:39 AM
Hi I want to extract a CSV file from SQL query one of the columns should be present in the CSV twice, diffirent position. I had the same error is there a workaround ? Thank you
... View more
08-30-2021
07:33 AM
Thank you so much. That was also the case with my NiFi instance. The replacement of this .jar package saved me lots of time.
... View more
02-28-2021
10:23 PM
1 Kudo
@murali2425 Here are two possible solutions. At solution 1 the value is set into an array. At solution 2 I took the flowfile-content (json) and set it into an attribute. Then you can work with expression language to get the value. For testing the syntax I recommend this site: http://jsonpath.herokuapp.com/
... View more
01-13-2021
04:47 AM
Did you already think about to transform your csv to json and then rebuild it with JOLT?
... View more
01-08-2021
05:23 AM
2 Kudos
@murali2425 The solution you are looking for is QueryRecord configured with a CSV Record Reader and Record Writer. You also have UpdateRecord and ConvertRecord which can use the Readers/Writers. This method is preferred over splitting the file and adds some nice functionality. This method allows you to provide a schema for both the inbound csv (reader) and the downstream csv (writer). Using QueryRecord you should be able to split the file, and set attribute of filename set to column1. At the end of the flow you should be able to leverage that filename attribute to resave the new file. You can find some specific examples and configuration screen shots here: https://community.cloudera.com/t5/Community-Articles/Running-SQL-on-FlowFiles-using-QueryRecord-Processor-Apache/ta-p/246671 If this answer resolves your issue or allows you to move forward, please choose to ACCEPT this solution and close this topic. If you have further dialogue on this topic please comment here or feel free to private message me. If you have new questions related to your Use Case please create separate topic and feel free to tag me in your post. Thanks, Steven
... View more
12-23-2020
09:33 AM
Hi Matt, Great, with your suggestion, I got what I was expecting. Thank You, --Murali
... View more
12-22-2020
10:31 PM
Hi, @murali2425 @vchhipa It seems some dependency issue while building your custom NiFi Processor, org.apache.nifi:nifi-standard-services-api-nar dependency needs to be added in pom.xml of nifi-*-nar module. Ref here <dependency>
<groupId>org.apache.nifi</groupId>
<artifactId>nifi-standard-services-api-nar</artifactId>
<version>1.11.3</version>
<type>nar</type>
</dependency> Please modify your pom.xml and rebuild and see whether that fixes the issue. Please accept the answer you found most useful.
... View more
05-18-2020
01:27 AM
The issue is with the missing of square brackets at starting and @ ending .. The working query is .. [{ "$group": { "_id": { "X": "$X", "Y_DT": "$Y_DT", "Z": "$Z" }, "adj": {"$sum": "$adj" }, "bjc": {"$sum": "$bjc" }, "jbc": {"$sum": "$jbc" }, "mnk": {"$sum": "$mnk"} } }]
... View more
05-04-2020
11:17 AM
Hi, i think your data-format .xlsx is Content of your Flowfile and you don´t have to convert it firtstly? So you can easy add a UpdateAttribute Processor and add the Dynamic-Property "filename" With this Property you overwrite the actually set Filename (you can see there is already an Attribute called "filename" from beginning). As Value you can set the name via Attribute and add the format there e. g. ${filename:append('.xlsx')} or just hard coded like file123.xlsx And after that you add a PutFile or FTP or whatever Processor and send it to the Path you set there.
... View more
04-28-2020
11:02 AM
it did work after adding '\t' to read_csv as 2nd arg.
... View more