Created 03-28-2018 03:31 AM
Hi,
i do not see any processors to connect to Google Big Query and execute queries.
is there a way where i can get results from google big query using NiFi.?
Created 03-28-2018 04:13 AM
Unfortunately, there is no specialized processor to connect to Google Big Query and execute queries. There has been some discussions about a set of new processors to support various Google Cloud services, but those processors are still to be planned into a release.
Until then you can use ExecuteScript processor.
Here is an example on how to write a script using Python: https://cloud.google.com/bigquery/create-simple-app-api#bigquery-simple-app-print-result-python . At https://cloud.google.com/bigquery/create-simple-app-api you can see other examples using other languages also supported by ExecuteScript processor.
Obviously, there is always the possibility to develop your own processor leveraging the Java example provided by Google doc. Example of how to build NiFi custom processor:
https://community.hortonworks.com/articles/4318/build-custom-nifi-processor.html
If this response addressed reasonably your question, please vote and accept answer.
Created 03-28-2018 01:32 PM
The community processor mentioned by Tim is a good example on how to write a custom processor. It is limited to Put action and quite old. You would have to rebuild it using more up-to-date libraries.
Community processors are not supported by Hortonworks.
Created 03-28-2018 01:23 PM
There is an open source community processor you can try:
Created 03-28-2018 09:13 PM
Can we use their REST APIs to download bigquery tables?
Hi @Timothy Spann , we cannot use any other processors which are not supported.
Regards,
Sai
Created 03-31-2018 12:10 AM
You can use their API to download the data from those tables. Those examples show how to select the data. However, you may deal with a lot of data. You may want to extract it from BigQuery, store it in Google Cloud Storage Bucket (GCS) and connect NiFi to GCS which is supported nicely with GCS processors to list, fetch, put, delete from GCS. That is the most efficient way.
Look at this reference to see how to extract the data: https://cloud.google.com/bigquery/docs/exporting-data
You can schedule a job to extract and put to GCS bucket and NiFi will just pick it up.
Created 04-02-2018 04:30 PM
My file names will change everyday like datafile_yyyymmdd.json , so we still need to create a scheduled job using Nifi and their RESTAPI to move the file to GCS.? Or is there a way that we can separate this .?
Regards,
Sai
Created 06-02-2018 09:37 AM
I had the same isssue and needed to create a processor on my own. Have a look at it:
https://datamater.io/2018/06/02/nifi-openaq-get-bigquery-processor/
Created 09-23-2018 01:47 PM
Are these processors public so others can benefit from them?
Created 10-17-2018 05:11 PM
Can't find it 😕 Where is the code source or the nar file please ?
Thanks for response.
,Hi @Pawel Leszczynski
Can't find it... Where is the code source of this processor ?
Thanks for your response.