Member since
09-29-2015
871
Posts
723
Kudos Received
255
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3367 | 12-03-2018 02:26 PM | |
2316 | 10-16-2018 01:37 PM | |
3633 | 10-03-2018 06:34 PM | |
2408 | 09-05-2018 07:44 PM | |
1834 | 09-05-2018 07:31 PM |
08-03-2016
04:10 PM
3 Kudos
ConsumeKafka and PublishKafka use the Kafka 0.9 client library. GetKafka and PutKafka use the Kafka 0.8 client library.
... View more
07-28-2016
04:57 PM
6 Kudos
The flow you see in the UI comes from the conf directory flow.xml.gz, unrelated to the H2 database. If you installed two NiFi instances you should have two completely separate directories like /opt/nifi1 and /opt/nifi2, each with their own conf sub-directories, each with their own flow.xml.gz. If you had an existing NiFi instance with a flow, and then copied that directory to nifi2, it would start with the same flow, but from there if you made changes to either they should be separate.
... View more
07-28-2016
12:53 AM
Glad to hear it is working. As far as the broker, NiFi is not doing anything special, it just takes the value entered in the processor config and passes it to Kafka as the property ProducerConfig.BOOTSTRAP_SERVERS_CONFIG (which is "bootstrap.servers"). So it really comes down to what Kafka does with this list. I would think it can be any hostname or IP that can be resolved.
... View more
07-28-2016
12:33 AM
4 Kudos
You can attach the debugger from your IDE to NiFI... In NiF's conf directory in bootstrap.conf there is a line commented out like this: #java.arg.debug=-agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=8000
If you uncomment that line and restart NiFi, the Java process will be listening for a debug connection on port 8000. If you want the process to wait for a connection before starting, you can also set suspend=y. Connecting your IDE debugger to the Java process will be specific to which IDE you are using. You can also add addition logging statements in your processor code using the logger return from getLogger() to see how far it is getting. The log-levels are controlled through the logback.xml file in the conf directory. The default level for processors is WARN: <loggername="org.apache.nifi.processors"level="WARN"/>
... View more
07-27-2016
08:07 PM
I just realized something... GetKafka and PutKafka use the kafka-client 0.8.2, and there are newer processors in 0.7.0 called ConsumeKafka and PublishKafka which use kafka-client 0.9.0.1. Since you are using Kafka 0.9 I think we should be using ConsumeKafka here. Lets see if that works any better, and sorry for the confusion.
... View more
07-27-2016
07:29 PM
Hi Stephanie, Were you able to try creating another GetKafka processor for the same topic on your NiFi graph to see if it experiences the same problem? The reason I wanted to try this was because each GetKafka processor has a property for Client Name which is specific to the processor's id, it ends up being something like "NiFi-<uuid>". This id will end up being stored in ZooKeeper somewhere to identify the consumer. So I wanted to see if creating a new GetKafka processor, and thus a new client id, would get the same problem or not.
... View more
07-27-2016
06:00 PM
Glad we got past the first problem... out of curiosity, if you created another new GetKafka processor pointing at the same Kafka topic, does it also get the same error? I was reading this old thread that seemed related, and it hinted that some kind of bad state might be stuck in ZooKeeper: http://apache-nifi-developer-list.39713.n7.nabble.com/GetKafka-blowing-up-with-assertion-error-in-Kafka-client-code-td9098.html
... View more
07-27-2016
01:06 PM
What do you want to do with the hashtags? If you want to get a new flow file for each hashtag you can use the SplitJson processor with a JSONPath value of $.twitter.hashtags
... View more
07-27-2016
02:48 AM
6 Kudos
As you mentioned, NiFi does offer many capabilities that can be used to perform ETL functionality. In general though, NiFi is more of a general purpose dataflow tool. NiFi clusters usually scale to dozens of nodes, so if your "transform" needs to run on 100s (or 1000s) of nodes in parallel, then it may be better to use NiFi to bring data to another processing framework like Storm, Spark, etc. Similar for the "extract" part... NiFi has capabilities to extract from relational databases with ExecuteSQL and QueryDatabaseTable processors which can solve many uses case, and for more extreme use-cases something like sqoop can leverage a much larger Hadoop cluster to perform the extraction. So as always there is no single correct answer and it depends a lot on the use-case.
... View more
07-27-2016
02:18 AM
It looks like there are two different connections being attempted here... the first one is to 52.90.171.224:2181 and then the one where the error is coming from is to localhost:2181. Can you confirm if 52.90.171.224 is a remote server, or is it the same IP of the machine where NiFi is running? Also, are you running a NiFi cluster or single instance? and if running a cluster, are you running an embedded ZooKeeper for NiFi's state management? Just trying to see if ZooKeeper is being used for anything else here.
... View more