Member since
11-16-2015
902
Posts
664
Kudos Received
249
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
148 | 09-30-2025 05:23 AM | |
617 | 06-26-2025 01:21 PM | |
452 | 06-19-2025 02:48 PM | |
696 | 05-30-2025 01:53 PM | |
9707 | 02-22-2024 12:38 PM |
12-14-2016
02:47 PM
Yes you can add a dynamic property whose value is a regular expression (see the documentation for more details).
... View more
12-14-2016
02:45 PM
The UI uses the REST API so you can do it programmatically the same way: /flowfile-queues/{id}/drop-requests
... View more
12-13-2016
02:44 PM
Yes, you can use something like the regex from step 2 above in a RouteOnContent processor, or after the ExtractText (step 2 above) you can use RouteOnAttribute looking for values of column.2.
... View more
12-08-2016
06:50 PM
The documents at the link above are for Apache NiFi 1.1.0, but HDF 2.0.0 was built with NiFi 1.0.0. The ability to append was added to NiFi 1.1.0 under NIFI-1322, so will likely be available in an upcoming version of HDF. The docs at that site are always for the latest version of Apache NiFi, it is recommended that you use the docs that come with your version of HDF/NiFi, using the Help option from the top-right hamburger menu in your running instance of HDF/NiFi.
... View more
12-08-2016
05:49 PM
Yeah we should probably trim that URL before using it, please feel free to write a Jira for that if you like.
... View more
12-08-2016
05:39 PM
1 Kudo
So if the customer_name value for id=CCCDD was "Matt" then you'd like the first output row to read: XXXXX, BBBBB, CCCCC, CCCDD, Matt Is that correct? If so, you could do the following: Use SplitText to split the incoming CSV into one flow file per line ExtractText to store the four column values as attributes (example template called Working_With_CSV here), let's assume the attribute for the fourth column is called "column.4" ReplaceText to set the content of the flow file to a SQL statement "select customer_name from table where id=${column.4} limit 1" ExecuteSQL to execute the statement ConvertAvroToJson to get the record into JSON (for further processing) EvaluateJsonPath to get the value of customer_name into an attribute (named "customer.name" with a JSON Path of $[0].customer_name or something like that) ReplaceText to set the row back to the original columns plus the new one, with something like "${column.1},${column.2},${column.3},${column.4}, ${customer.name}" (optional) MergeContent to join the rows back together (if you need them as one file)
... View more
12-08-2016
05:01 PM
Looks like your connect URL has a space as the first character? For the second URL, that might be one of two things. The first (less likely) is that you want your client to use hostnames to resolve name nodes (see my answer here). However I would've expected an error message like the one in that question, not the one you're seeing.
I think for your second URL issue the problem is that Apache NiFi (specifically the Hive processors) doesn't necessarily work with HDP 2.5 out of the box, because Apache NiFi ships with Apache components (such as Hive 1.2.1), whereas HDP 2.5 has slightly different versions. I would try Hortonworks Data Flow (HDF) rather than Apache NiFi, as the former ships with HDP versions for Hive.
... View more
12-08-2016
04:11 PM
1 Kudo
Your original URL "jdbc://hive2://localhost:10000/default" has slashes between jdbc: and hive2, this should instead be jdbc:hive2://localhost:10000/default For the zookeeper version of the URL, that is a known issue (NIFI-2575), I would recommend correcting the original URL and using that (port 10000 should be opened/forwarded on the sandbox already).
... View more
12-07-2016
09:56 PM
There is a table for JDBC type values here, for your data types they are as follows: Type Value INTEGER 4 TIMESTAMP 93 DOUBLE 8 LONGVARCHAR -1
... View more
12-06-2016
09:31 PM
Each processor is responsible for reading and writing whichever attributes it wants to for the purposes of its processing, and those attributes are available in each processor's documentation. SplitJson for example writes the following attributes to each output flow file: Name Description fragment.identifier All split FlowFiles produced from the same parent FlowFile will have the same randomly generated UUID added for this attribute fragment.index A one-up number that indicates the ordering of the split FlowFiles that were created from a single parent FlowFile fragment.count The number of split FlowFiles generated from the parent FlowFile segment.original.filename The filename of the parent FlowFile These were added to NiFi 1.0.0 (HDF 2.0) under NIFI-2632, so if you are using a version of NiFi/HDF before that, that's why you won't see these attributes populated by SplitJson.
... View more