Member since
07-30-2019
3423
Posts
1630
Kudos Received
1010
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 242 | 01-13-2026 11:14 AM | |
| 462 | 01-09-2026 06:58 AM | |
| 666 | 12-17-2025 05:55 AM | |
| 727 | 12-15-2025 01:29 PM | |
| 605 | 12-15-2025 06:50 AM |
01-19-2017
09:26 PM
step1 : remove the authorization.xml from all the nodes rm /var/lib/nifi/conf/authorizations.xml step2 : update node identities in advanced_nifi_ambari_ssl_configuration section with <property name="Node Identity 1">CN=hostname1, OU=XXXXX</property> ... for all the nodes (make sure you remove comment tags in XML). step3 : restart the NiFi service found that authorizations.xml is not being update after first generation, hence this is causing the problem.
... View more
10-12-2016
01:51 PM
Hi @mclark, Thanks for the reply and for time you spend on zoom. It was confusing that when i make a change in the component it gets added to the Parent group if i don't override, which was some kind of reverse inheritance. The message said only below which made me think there is something wrong: "Showing effective policy inherited from Process Group Group1.
Override this policy." Thanks Again for clarifying !!
... View more
03-02-2017
03:38 PM
Hi i have 5 separate queues for 5 different processors, everytime i'm going to each processor and clearing the each queue its taking me lot of time, is there any way to clear all the queue's at same time ? please help me with this thanks Ravi
... View more
09-21-2016
05:29 PM
1 Kudo
There is the possibility that the time could differ slightly (ms) between when both now() functions are called in that expression language which could cause the result to push pack to 11:58:59. To avoid this you can simply reduce 43260000 by a few milliseconds (43259990) to ensure that does not happen so 11:59:00 is always returned.
... View more
09-21-2016
01:24 PM
OMG, stupid me 😄 Thanks @mclark , exactly that solved the issue, sorry for bothering
... View more
09-19-2016
08:28 PM
1 Kudo
@mclark Wow. Thanks. This may be the direction we're looking for. Thank you. This will certainly help. I feel some additional kafka questions coming along however: particularly on the topic of linking ConsumeKafka with GetKafka and it's properties, but this is definitely a big leap in where we want to be.
... View more
10-12-2016
02:24 PM
1 Kudo
@Saikrishna Tarapareddy The FlowFile repo will never get close to 1.2 TB in size. That is a lot of wasted money on hardware. You should inquire with your vendor about having them split that Raid in to multiple logical volumes, so you can allocate a large portion of it to other things. Logical Volumes is also a safe way to protect your RAID1 where you OS lives. If some error condition should occur that results in a lot of logging, the application logs may eat up all your disk space affecting you OS. With logical volumes you can protect your root disk. If not possible, I would recommend changing you setup to a a bunch of RAID1 setups. With 16 x 600 GB hard drives you have allocated above, you could create 8 RAID1 disk arrays. - 1 for root + software install + database repo + logs (need to make sure you have some monitioring setup to monitor disk usage on this RAID if logical volumes can not be supported) - 1 for flowfile repo - 3 for content repo - 3 for provenance repo Thanks, Matt
... View more
09-13-2016
06:22 PM
Thank you! What I missed here is to grant permission "view the data" of nifi flow to nodes in cluster.
... View more
09-08-2016
03:59 AM
Yep what you describe with UpdateAttribute/MergeContent sounds perfectly fine. What you'll want there precisely will depend on how many relationships you have out of RouteText. As for concurrent tasks I'd say it would be 1 for GetFile 1 for SplitFile 2...4 or 5 or so on RouteText. No need to go too high generally. 1 for MergeContent 1 to 2 for PutHDFS You don't have to stress too much on those numbers out of the gate. You can run it with minimal threads first, find any bottlenecks and increase if necessary.
... View more