Member since
05-20-2016
13
Posts
4
Kudos Received
0
Solutions
03-21-2017
08:27 PM
9 Kudos
@Faruk Berksoz - Below query should work fine: SELECT * FROM ( SELECT * FROM TB.TABLE1 ) T;
... View more
02-27-2017
11:41 PM
1 Kudo
Hello @Faruk Berksoz In the screenshot, TailFile Task/Time is shown as 30 (times) in the last 5 min (300 secs). If you scheduled TailFile to run every 10 sec, I think the stats is correct. It seems TailFile is scheduled correctly but there hasn't been new lines found so no FlowFile is produced by TailFile. When new lines are added to the file that being watched by the TailFile, those will be picked up by TailFile and passed to PublishKafka. If you're sure that new lines are appended but not seeing any data is ingested to NiFi, then please elaborate the issue. Regards, Koji
... View more
02-24-2017
05:45 PM
2 Kudos
@Faruk Berksoz Kafka - YES for all scenarios. Kafka is not for storing. Kafka is for transport. Your data still needs to land somewhere, e.g. As you mentioned that is HBase via Phoenix, but it could also be HDFS or Hive. 1. Yes. Flume is ok for ingest, but you still need something else to post to Kafka (Kafka Producer), e.g. KafkaConnect. 2. No. Spark Streaming is appropriate for consumer applications, not really for your use case which is about ingest and post to Kafka. 3. No. Same response as for #2 4. No. Storm is for appropriate for consumer applications, not really for your use case which is about ingest and post to Kafka. 5. Could work, not recommended. The most common architectures are: a) Flume-> KafkaConnect-> Kafka; consumer applications are built using either Storm or Spark Streaming. Other options are available, but less used. b) Nifi -> Kafka -> Storm; consumer applications are built using Storm; this is Hortonworks DataFlow stack c) Others (Attunity, Syncsort) -> Kafka -> consumer applications built in Storm or Spark Streaming Since I am biased, I would say go with b) - Storm or Spark Streaming, or both. I'm saying that only because I am biased but because each of the components scale amazingly and because I used Flume before and don't want to go back there once I've seen what I can achieve with NiFi. Additionally, HDF will evolve an integrated platform for stream analytics with visual definition of flows and analytics requiring the least programming. You will be amazed of the functionality provided out of box and via visual definition and that is only months away. Flume is less and less used. NiFi does what Flume does and much beyond. With NiFi writing the producers to Kafka is trivial. Think beyond your current use case. What other use cases can this enable?... One more thing. For landing data to HBase you can still use NiFi and its Phoenix connector to HBase. Another scalable approach.
... View more
08-24-2017
12:35 PM
Hello Josh, I am trying to connect to apache phoenix v4.7.0.2.5.6.0-40 using JDBC in node.js. According to the comment in the bellow question link asked by me, I will be needing thick phoenix client jar for this. Is there any way I can get this jar. https://stackoverflow.com/questions/45837004/issue-in-connecting-apache-phoenix-with-node-js-using-npm-jdbc-package?noredirect=1#comment78677574_45837004 Any reference will be really helpful.
... View more