Created 03-09-2018 06:49 AM
I have a special case which is required to write out Parquet file into local directory.
I tried not to enter 'Hadoop Configuration Resources' but it will instantly throw exceptions.
If I entered "Hadoop Configuration Resources", then the output path is actually points to path inside the HDFS cluster.
So, I would like to know, Does PutParquet processor support writing to Local Folder?
Created 03-09-2018 02:34 PM
I think you can provide a core-site.xml that refers to a local filesystem as the default FS rather than an HDFS file system by setting the fs.default.name property to file:/// (see the answer on this SO post for more details).
Created 03-09-2018 02:34 PM
I think you can provide a core-site.xml that refers to a local filesystem as the default FS rather than an HDFS file system by setting the fs.default.name property to file:/// (see the answer on this SO post for more details).
Created 03-09-2018 03:29 PM
Matt is correct, and just to elaborate further... the Parquet API only allows writing through the Hadoop FileSystem, so the only way NiFi can write Parquet to local filesystem is through the configuration Matt mentioned above.
Created 03-12-2018 07:03 AM
Thanks all. One more question: How does PutParquet Processor connect to HDFS? By WebHDFS , HttpFS or Native ?
Created 03-12-2018 01:04 PM
Native through the standard Hadoop Java client.