Member since
06-08-2017
1049
Posts
518
Kudos Received
312
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 11124 | 04-15-2020 05:01 PM | |
| 7022 | 10-15-2019 08:12 PM | |
| 3068 | 10-12-2019 08:29 PM | |
| 11251 | 09-21-2019 10:04 AM | |
| 4190 | 09-19-2019 07:11 AM |
08-20-2018
01:53 PM
1 Kudo
@CHEH YIH
LIM
I think you are using extract text processor to extract the content and keep as attribute to the flowfile if yes then change Maximum Buffer Size 1 MB Specifies the maximum amount of data to buffer (per file) in order to apply the regular expressions. Files larger than the specified maximum will not be fully evaluated. Maximum Capture Group Length 1024 Specifies the maximum number of characters a given capture group value can have. Any characters beyond the max will be truncated. These two property values as per your flowfile size. - If the Answer helped to resolve your issue, Click on Accept button below to accept the answer, That would be great help to Community users to find solution quickly for these kind of issues.
... View more
08-20-2018
04:15 PM
3 Kudos
hive.merge.cardinality.check=false is a bad idea. The logic controlled by this property checks if the ON clause of your Merge statement is such that more than 1 row from source side matches the same row from target side (which only happens in WHEN MATCHED clause). Logically what this means is that the query is asking the system to update 1 existing row in target in 2 (or more) different ways. This check is actually part of SQL standard definition of how Merge should work. You either need examine your data or the ON clause but disabling this check, when it throws a cardinality_violation error, may lead to data corruption later.
... View more
08-18-2018
11:40 PM
Thanks very much @Shu for your solution . I managed to find a way which is almost similar to yours.
... View more
08-15-2018
04:05 AM
@Shu Please find attached the list queue after GetFile. As soon as i start my UpdateRecord the List queue become blank. attaching the listqueue after UpdateRecord processor Expected output is to concatenate the row 9th value (FEBetaPack01) and 23rd row value (FE_ChargerSim_CurrentRipple)under column 'filename ' as (FEBetaPack01-FE_ChargerSim_CurrentRipple). Please help. Thanks
... View more
08-09-2018
02:33 PM
1 Kudo
@Rinki Flow: 1. You can list out the files from the directory on every first day of month and check the filename attribute using RouteOnAttribute Processor to get only the current date files. In RouteOnAttribute processor you can use either of the above attributes to making use of nifi expression language we can only filtering out only the required files. 2.You can use ReplaceText processor to replace all this required metadata and store into HDFS/Hive..etc i'm thinking filetype is csv,avro,json so i kept expression like ${filename:substringAfter('.')} Replacement Value
${filename},${file.creationTime},${filename:substringAfter('.')},${file.size} To store the data to table you can use PutHDFS and create table on top of this directory. 3.You can use cron schedule to run the processor on first day of month and Execution in only on Primary node - If the Answer helped to resolve your issue, Click on Accept button below to accept the answer, That would be great help to Community users to find solution quickly for these kind of issues.
... View more
08-09-2018
01:38 PM
Thanks @Shu
... View more
08-08-2018
03:39 PM
@Bharath
Sudharsanam
Hmm, If possible please share your flow template (or) all the processors that you are using and configs screenshots of them.
... View more
08-08-2018
02:22 AM
@Bharath
Sudharsanam
Yes, leave the Fetch File processor property as default. processor will fetches those files from the directory.
... View more
01-09-2019
09:16 AM
Any progress on getting MoveHDFS to accept attributes in Output Directory? It seems difficult not be able to have a dynamic solution as mentioned in this thread.
... View more