Member since
06-08-2017
1049
Posts
518
Kudos Received
312
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 11245 | 04-15-2020 05:01 PM | |
| 7156 | 10-15-2019 08:12 PM | |
| 3134 | 10-12-2019 08:29 PM | |
| 11565 | 09-21-2019 10:04 AM | |
| 4361 | 09-19-2019 07:11 AM |
08-07-2018
10:03 PM
@Bharath
Sudharsanam
Fetch File processor works in conjunction with List file processor as list file processor adds ${absolute.path,${filename} attributes to the flowfiles then Fetch File processor fetches those files. If you are not using ListFile processor then you need to having fully qualified file path to be mentioned in the File to Fetch property value and this processor needs an upstream connection to trigger. Example: I want to fetch file_fetch.txt file from temp directory then Fetch File configs: Now FetchFile processor will fetches this file_fetch.txt from temp directory and transfers to success relationship. If you are having file_path attribute with value as /temp/file_fetch.txt associated with the flowfile then Configs: Now the processor checks the attribute value and dynamically fetches the files from the directory if NiFi have access to those directories. Flow: As this processor expects upstream connection and i'm not using ListFile processor instead i use GenerateFLowfile (added file_path attiribute) processor to trigger Fetch File processor (or) you can simply use ListFile + FetchFile processors also.
... View more
08-06-2018
01:00 PM
@Frederic
Le Texier
I'm looking into NiFi-1.7.1 documentation and the source code link of MoveHDFS processor(which have updated documentation) https://github.com/apache/nifi/blob/master/nifi-nar-bundles/nifi-hadoop-bundle/nifi-hdfs-processors/src/main/java/org/apache/nifi/processors/hadoop/MoveHDFS.java#L150 In apache nifi org the documentation is created based NiFi-1.5 version, probably these missing things are updated NiFi 1.5+ versions. As you have proposed below, we need to use RouteOnAttribute processor to make decision based on the attribute.
... View more
08-05-2018
01:15 AM
@Julio
Gazeta
As you are having Mode property value as Update and also Change the Upsert property value to true //When true, inserts a document if no document matches the update query criteria; this property is valid only when using update mode, otherwise it is ignored For more details regards to updating the docs refer to this and this links.
... View more
08-03-2018
07:00 PM
@Frederic
Le Texier
The issue is with the attribute ${Librairie} in Output Directory property value. As per NiFi-1.7 MoveHDFS processor Usage OutputDirectory Property value supports Expression Language but evaluated using variable registry only, But for InputDirectory/File property value will be evaluated using flowfile attributes and variable registry. To resolve this issue: Define your ${Librairie} attribute value in variable registry if possible. As right now you are having the variable as flowfile attribute which this value will not be evaluated and files are moving to ${path_to_prod_AS} directory.
... View more
08-02-2018
10:14 PM
@Danilo Sousa if you want to execute sequentially then you can use Success of PutHiveQL processor to trigger another job(i.e. start table B). Flow: 1.GenerateFlowfile //start with tableA 2.PutHiveQL 3.ReplaceText //to prepare tableB statement
4.PutHiveQL
... View more
08-01-2018
10:36 PM
1 Kudo
@Nikil Katturi You can use filter function on the data frame import org.apache.spark.sql.functions._
df.filter(col("<timestamp_column_name>") > lit(date_sub(current_date,1))).show(10,false)
... View more
07-31-2018
11:41 PM
1 Kudo
@Danilo Sousa If you want to move data between two hive tables then you don't need to use SelectHiveQL processor at all. You create hive statement like below insert into <db_name>.<final_table> select * from <db_name>.<rawdata> Then execute the above statement using PutHiveQL processor. To incrementally run this process then you need to store the state i.e. until what time you have already processed the data from rawdata table. Then only select the new data after the state value. Please refer to this and this link for more details how to incrementally copy data in hive.
... View more
07-25-2018
01:21 PM
1 Kudo
@Mohammad
Soori
Connect only splits relationship from Split Text to Publish kafka processor Auto terminate original relationship Go to configure --> settings tab--> check the original box and regarding failure relationship connect to putEmail processor to get notified if some thing went wrong. Flow:-
... View more
07-25-2018
01:03 PM
@Sudheer K We need to configure demarcator as newline (i.e shift+enter or \n) Merge Content Configs:
... View more
07-25-2018
12:51 AM
2 Kudos
@Andrew Riffle Use ModifyBytes processor which will drop the contents of your flowfile. Set the property value remove all content to true Processor configs:-
... View more