Created 07-18-2016 08:54 PM
Goal is to read the data from RDBMS table and store it in HDFS in avro format with additional column , let say source table has 5 columns , as part of the ingestion I would like to add additional column say "ingest_datetime" with current_time value before nifi stores the file in HDFS finally HDFS should have avro file with 6 columns in the end .
Currently I am using ExecuteSQL --> PutHDFS processors
Created 07-18-2016 09:06 PM
Currently there aren't any processors that perform direct manipulation of Avro, although we definitely would like to have some.
Possible options to work around this...
Happy to help answer any questions if going the custom Java processor route.
Created 07-18-2016 09:06 PM
Currently there aren't any processors that perform direct manipulation of Avro, although we definitely would like to have some.
Possible options to work around this...
Happy to help answer any questions if going the custom Java processor route.
Created 07-19-2016 06:49 PM
Thanks for the response ..
Do you know when the new JOLT transform processor is going to be releasing ? existing 0.61. or 0.7 does not have this new processor you are talking about, but NIFI-361 ticket is talking about it.
Created 07-19-2016 07:00 PM
It is in the 0.7.0 release, part of the stanadard bundle:
Created 07-19-2016 01:51 PM
@Sreekanth Munigati , As @Bryan Bende mentioned, there is no direct way of manipulating Avro data, but in your case you can try modifying SQL being executed by ExecuteSQL processor to add an additional column in SQL itself.