Member since
06-08-2017
1049
Posts
518
Kudos Received
312
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 11986 | 04-15-2020 05:01 PM | |
| 7952 | 10-15-2019 08:12 PM | |
| 3593 | 10-12-2019 08:29 PM | |
| 12979 | 09-21-2019 10:04 AM | |
| 4845 | 09-19-2019 07:11 AM |
05-24-2019
04:36 AM
@Yasuhiro Shindo Container killed exit code most of the time is due to memory overhead If you haven't specified spark.yarn.driver.memoryOverhead or spark.yarn.executor.memoryOverhead these params in your spark submit then add these params (or) if you have specified then increase the already configured value. Please refer to this link to decide overhead value.
... View more
05-24-2019
03:49 AM
1 Kudo
@Alvarez Rafa QueryDatabaseTable processor stores the state when processor ran for the first time based on Max value column(idmovil). For the next run processor only pulls the changes from the table based on idMovil column. You can check the last state value by RightClick on processor -> View State To clear the state stop the processor and RightClick on processor -> View State and then clear state. Once we clear the state then processor pulls all records from the table. --- If you are not facing this issue then please attach your flow screenshot and scheduling on QueryDatabaseTable processor.
... View more
05-08-2019
02:39 AM
@HanYan Tan Could you look into zeppelin-4140 jira is reported for the same issue. Checkout the comments associated with the jira for more details, as mentioned in jira comments "jdbc interpreter Binding Mode Per User Scoped and isolated mode per user the temporary table is dropped" Try with the above setting and check does the issue resolved or not 🙂
... View more
05-02-2019
12:27 AM
1 Kudo
@Raj Negi Use nifi expression language to replace function to replace "[ and ]" with [ and ] ${'$1':unescapeJson():replace('"[','['):replace(']"',']')}
... View more
05-01-2019
12:06 AM
1 Kudo
@Raj Negi After FetchHbaseRow processor use ReplaceText processor with below configs: Search Value (?s)(^.*$) Replacement Value ${'$1':unescapeJson()} //capture all the data and apply nifi expression language unescapeJson function. Character Set UTF-8 Maximum Buffer Size 1 MB //change as per your flowfile size Replacement Strategy Regex Replace Evaluation Mode Entire text Flow: 1.FetchHbaseRow
2.ReplaceText
--other processors
... View more
04-09-2019
03:16 AM
1 Kudo
@Kevin Lahey Not sure if you are using NiFi cluster (or) not, could you try to run ListS3 processor only on Primary Node only. As this processor intended to run only on primary node as per documentation.
... View more
03-22-2019
01:09 AM
@sri chaturvedi Instead of using UpdateAttribute processor's state use DistributedMapCache and you can fetch the stored value across the cluster. Use PutDistributedMapCache processor to store the value that got assigned recently then use FetchDistributedMapCache processor to Fetch the store value then apply your logic(increment..etc) to assign new value then overwrite the already stored value in DistributedMapCache using PutDistributedMapCache processor. Use this and this links as references for configuring Distributedmapcache processors/controllers.
... View more
01-23-2019
01:52 AM
@john y We can directly access filesize with ${fileSize} and this attribute expression will result the actual filesize value of flowfile.
... View more
01-18-2019
02:43 AM
@Manish Parab
Sure, In NiFi processors that triggers the flow(scheduled to run in cron) we need to run the processors on primary node only and running on all nodes means we are triggering n times the same processor on each node. That means NiFi each node works with data specifically that receives, in case of Getmongo processor(triggers the flow in this case) when running on all nodes will pull same data. - Run GetMongo(source processor) to run on primary node then distribute the load using RemoteProcessorGroups (or) connectionloadbalancing across the cluster.
... View more
01-18-2019
12:32 AM
@Manish Parab Your GetMongo processor is running on All nodes that means same data is pulled on all nodes.. If you select EvaluateJsonPath processor to run on Only primary node then all the other nodes flowfiles will be left queue before EvaluateJsonPath processor, Because you are not processing flowfiles that are pulled from all other nodes except of PrimaryNode. Run GetMongoProcessor only on primary node and keep EvaluateJsonPath Processor to run on all nodes, Reason to keep EvaluateJsonPath processor on all nodes if NiFi primary node changed then EvaluateJsonPath processor not going to processor the flowfile that are listed on old PrimaryNode.
... View more
- « Previous
- Next »