Member since
09-02-2016
56
Posts
6
Kudos Received
0
Solutions
02-27-2018
12:24 PM
Hi, Since we can mention only one prefix in ListS3 processor I am trying to access AWS S3 using Python boto3 in NiFi ExecuteScript processor. If this succeeds, I can send a list of folder paths to the python script to get files from various folders under S3 bucket. I provided the path of boto3-1.6.0.tar.gz in ModuleDirectory but I get the below error. @Matt Burgess @Bryan Bende @Matt Foley Please let me know how to fix this and achieve use case. ExecuteScript Processor : Error: ExecuteScript[id=d6cf51e8-0161-1000-32de-7748af781842] Failed to process session due to org.apache.nifi.processor.exception.ProcessException: javax.script.ScriptException: ImportError: No module named boto3 in <script> at line number 1: javax.script.ScriptException: ImportError: No module named boto3 in <script> at line number 1
... View more
Labels:
- Labels:
-
Apache NiFi
02-23-2018
10:15 AM
Hi @Andrew Lim Thanks for the article. I tried to execute this using NiFi 1.5 but I get an error "already in use for an active callback or an InputStream created by ProcessSession.read(FlowFile) has not been closed" in QueryRecord processor. PFb excerpt from logs - ErrorHandlingInputStream[FlowFile=StandardFlowFileRecord[uuid=9688750b-0f83-4c64-91c6-80499ef63c24,claim=StandardContentClaim [resourceClaim=StandardResourceClaim[id=1519378477705-1, container=default, section=1], offset=532, length=179],offset=0,name=input.csv,size=179]] for StandardFlowFileRecord[uuid=9688750b-0f83-4c64-91c6-80499ef63c24,claim=StandardContentClaim [resourceClaim=StandardResourceClaim[id=1519378477705-1, container=default, section=1], offset=532, length=179],offset=0,name=input.csv,size=179] because the session was rolled back without the input stream being closed.
2018-02-23 15:10:17,160 ERROR [Timer-Driven Process Thread-3] o.a.nifi.processors.standard.QueryRecord QueryRecord[id=aff8e1fe-ff3c-3d77-9fcf-dcd3c1cf6ed9] QueryRecord[id=aff8e1fe-ff3c-3d77-9fcf-dcd3c1cf6ed9] failed to process session due to java.lang.IllegalStateException: StandardFlowFileRecord[uuid=9688750b-0f83-4c64-91c6-80499ef63c24,claim=StandardContentClaim [resourceClaim=StandardResourceClaim[id=1519378477705-1, container=default, section=1], offset=532, length=179],offset=0,name=input.csv,size=179] already in use for an active callback or an InputStream created by ProcessSession.read(FlowFile) has not been closed: {}
java.lang.IllegalStateException: StandardFlowFileRecord[uuid=9688750b-0f83-4c64-91c6-80499ef63c24,claim=StandardContentClaim [resourceClaim=StandardResourceClaim[id=1519378477705-1, container=default, section=1], offset=532, length=179],offset=0,name=input.csv,size=179] already in use for an active callback or an InputStream created by ProcessSession.read(FlowFile) has not been closed<br> Can you please tell what's wrong?
... View more
02-22-2018
06:24 PM
Hi @Bryan Bende How to get more than one column value from lookup file? PFB details. InputFile ID, Name
1,Elon
2,Mark
3,Jeff Lookup File ID,Company,Domain,HQ
1,Tesla,Automobile,Palo Alto
2,Facebook,SocialMedia,Menlo Park
3,Amazon,ECommerce,Seattle Output ID,Name,Company,Domain,HQ
1,Elon,Tesla,Automobile,Palo Alto
2,Mark,Facebook,SocialMedia,Menlo Park
3,Jeff,Amazon,ECommerce,Seattle Result RecordPath allows only one value. Any tips to get this kind of output?
... View more
02-22-2018
06:04 PM
Thanks for replying @Abdelkrim Hadjidj @Matt Burgess. Now I have two different schemas for input and output CSV for LookupRecord. I modified Result RecordPath to /COMPANY. When I start LookupRecord it takes in 6 flow files but doesn't return anything.
... View more
02-22-2018
04:45 PM
Hi There, I am trying to lookup records from a CSV file using LookupRecord and SimpleCsvFileLookupService but I am unable to get the final record with both values. Please find below details. Input File ID, Name
1,Elon
2,Mark
3,Jeff Lookup File ID,Company
1,Tesla
2,Facebook
3,Amazon Expected Output ID,Name,Company
1,Elon,Tesla
2,Mark,Facebook
3,Jeff,Amazon AvroSchema {
"name": "ceodata",
"namespace": "nifi",
"type": "record",
"fields": [
{ "name": "ID", "type": "string" },
{ "name": "NAME", "type": "string" },
{ "name": "COMPANY", "type": "string" }
]
} LookupRecord Processor Config CSVReader Config CSVWriter Config LookupFromCSVFileService Please let me know what's wrong in these configurations. @jfrazee @Bryan Bende @Abdelkrim Hadjidj @Timothy Spann @Aldrin Piri
... View more
Labels:
- Labels:
-
Apache NiFi
02-09-2018
01:54 PM
1 Kudo
Hi Andrew, Is there any way to automate clearing state from ListS3 processor when the flow is scheduled? I cannot clear the state every time I run the flow.
... View more
12-21-2017
10:22 AM
In a NiFi Cluster If a node goes down, it is known that the flow can be directed to another node, but what about the data that is queued in the failed node? Will have to wait until the node comes back up or it is taken care by another node? Please explain how it handles this scenario.
... View more
Labels:
- Labels:
-
Apache NiFi
10-12-2017
11:48 AM
Hi slachterman thanks for the post. I get below error while running PutHDFS processor. I didn't find the SNAPSHOT jars in HDInsight, May I know the HDI version you used? Can you help me fixing this error? Caused by: java.lang.NoSuchMethodError: org.apache.hadoop.security.ProviderUtils.excludeIncompatibleCredentialProviders(Lorg/apache/hadoop/conf/Configuration;Ljava/lang/Class;)Lorg/apache/hadoop/conf/Configuration; Thanks Vinay
... View more
10-12-2017
02:55 AM
Hello Team, Is it possible to ingest data to Azure SQL DB directly using NiFi? Thanks
... View more
Labels:
- Labels:
-
Apache NiFi
05-02-2017
12:13 PM
Thanks, @Wynner.
We are going with AWS cloud instance to setup NiFi Cloud, I wanted to know any specifications required for the cloud server. Thanks.
... View more