Member since
06-19-2017
62
Posts
1
Kudos Received
7
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2522 | 03-17-2022 10:37 AM | |
1758 | 12-10-2021 04:25 AM | |
2324 | 08-18-2021 02:20 PM | |
6278 | 07-16-2021 08:41 AM | |
1153 | 07-13-2021 07:03 AM |
07-13-2021
07:03 AM
We have solved this with help of wait and notify processor as we routed .hql file to puthql which interns routes success/failure to Notify . The Wait processor wait signal will release the .csv file to put into HDFS once Notify signal comes from Notify processor.
... View more
07-13-2021
06:52 AM
Hi Deepika, I assume that you are using ConsumeKafka OR ConsumeKafka2.0 version NiFi processor ,when you select option in SSL Context Service as StandardSSLContextService then you have to select right arrow as indicated below image . Properties of StandardSSLContextService will be prompted to you and you can enter the values of the properties Keystore Filename,Keystore Password,Key Password,Truststore Filename,Truststore Type. etc. After providing the values of the SSL properties , enable the Controller Service and start running the ConsumeKafka processor. ConsumeKafka2.0 processor properties Please update if above steps expected and works. Thanks
... View more
07-01-2021
12:45 PM
Hi , The InvokeHTTP processor provides couple of the write attributes with value for example invokehttp.status.code 401 invokehttp.status.message Unauthorized when you have this attributes in Failure or No Retry relationship ,you can use Replace text processor as below to overwrite the original flow into new flow file so that you send that in email . ReplaceText processor properties If you are looking for lot more message of the response body , please check whether you can configure in any attribute in InvokeHTTP processor so that you can use it ReplaceText processor to overwrite the original flow file
... View more
07-01-2021
04:22 AM
Hi, We are processing ZIP file contains multiple timestamp files (.hiveql,.csv) in distributed manner . We check the file extension whether it is .hql or .csv then we route the file to execute it PutHiveQL and PutHDFS processor respectively. The files(timestamp order starts with for example t1 or system timestamp) below contains in ZIP file to be extracted and processed in order. table_info.zip table_info_t1.hql
table_info_t1_1.csv
table_info_t1_2.csv
table_info_t2.hql
table_info_t2_1.csv
table_info_t2_2.csv
table_info_latest.hql
table_info_latest.csv Please find the below NiFi flow and RouteonAttribute property Is there any way to make us to wait first puthivesql executes first and give indication to putHDFS execution next for each timestamp file one by one order. Can we group each timestamp files into group and process the .hql file and the put .csv file into HDFS? @Nifi
... View more
Labels:
- Labels:
-
Apache NiFi
06-11-2021
03:16 AM
Hi , Thanks for the reply . As i mentioned that i have run the insert ino tableA select * from tableB (Assume Delta calculation query). Yes , we have managed table in hive and hive version is 3.0 where ACID are enabled by default .We have created the Hive managed table through HUE and executed the delta query via NiFi PutHQL processor and we got the error intermittently .
... View more
05-17-2021
04:16 AM
Hi , Thanks for the information . We have added this single property set hive.txn.stats.enabled=false and still we were getting the issue intermittenly from Nifi. My Solution architect found the cause for the issue after investigating Hive github code I believe. So we added to these two property below set hive.txn.stats.enabled=false set hive.stats.autogather=false The error disappeared and never come again .I would like to understand how to relate the hive github with those properties or how do have to troubleshoot this kind of strange issue .
... View more
04-22-2021
01:58 AM
Hi All, We are getting the below error intermittently for some tables and we have hive 3.1 . I am passing a sample query like insert into table test as select * from some table in Nifi PutHiveSQL processor . Hive target table is managed table and stored as avro format Error : PutHive3QL[id=d1652e76-54f7-30cf-8dc8-b1934cee3c26] Failed to update Hive for StandardFlowFileRecord[uuid=7447e7f0-4d1e-40fa-9965-5c98f7d11341,claim=StandardContentClaim [resourceClaim=StandardResourceClaim[id=1618924236521-4, container=default, section=4], offset=154416, length=3631],offset=0,name=20210421_test_branch_info,size=3631] due to java.sql.SQLException: Error while compiling statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.StatsTask. MetaException(message:Cannot change stats state for a transactional table ra_test.test_branch_info without providing the transactional write state for verification (new write ID -1, valid write IDs null; current state {"BASIC_STATS":"true","COLUMN_STATS":{}; new state {"BASIC_STATS":"true","COLUMN_STATS":{}); it is possible that retrying the operation will succeed, so routing to retry: java.sql.SQLException: Error while compiling statement: FAILED: Execution Error, Option tried : if we rerun the query for the same table through Nifi PutHiveSQL , then Nifi processor executes the query fine . sometimes it fails in the first run. Hive version : 3.0 and Nifi 1.11.4 Please assist what option to be checked in Hive side on this error. Thanks
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache NiFi
02-11-2021
03:29 AM
@MattWho In continuation of above logic and i implemented and it is working fine . I have three nodes nifi cluster where a,b,c .Here a node is primary node always . If a goes out of space , we will unzip the file from another node and distributes to b,c nodes and we push the file to HDFS folder and we capture success and failure message( in Updateattribute processor) as mentioned in the posts above . After that in mergerecord process , the queue is waiting for sometime and merging happens after 10mins around . I have a dependency of the merging processor as i have to share the status of flowfile to another API which is requesting for it. How we can speed up the MergeRecord processor here . If all three nodes have good amount of memory and the flowfiles pushed to HDFS and status updates (through UpdateProcessorattribute) and merging happens very quickly as expected in single node. But we have nodes(b,c) and 'a' primary node is out due to space issue in device .The processor unpack the .zip file then distribute to nodes tp put into HDFS .As we capture each success/failure status of PutHDFS (using updateattribute processor) and mergerecord is taking sometime to merge the status . Is this because of other two nodes processing it slowly or some other reason for that. I have attached screenshot of the partial flow .
... View more
01-25-2021
04:20 AM
Hi All, I am trying to build JSON string using Ifelse condition in ReplaText processor but i am getting error as invalid Invalid Expression . Working Ifelse condition : ${status:contains('0'):ifElse('Success','Failed')} This below string is not working and it is throwing invalid expression . Please assist on this . ${status:contains('0'):ifElse('{ "filename":"${overall_filename}", "status":"${status}", "message":"${overall_message}" }', '{ "filename":"${overall_filename}", "status":"${exc_status}", "message":"${overall_message}" }' ) } Thanks
... View more
Labels:
- Labels:
-
Apache NiFi
01-09-2021
09:30 PM
My input JSON will be like below [ { "filename":"test1.csv", "message":"HiveSQL query executed Sucessfull!", "status":"Success" }, { "filename":"test2.csv", "message":"HiveSQL query executed Sucessfull!", "status":"Success" }, { "filename":"test3.csv", "message":"HiveSQL query executed Sucessfull!", "status":"Success" } ]
... View more