Member since
09-23-2015
8
Posts
5
Kudos Received
5
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3009 | 02-11-2016 08:16 PM | |
885 | 01-11-2016 09:22 PM | |
1477 | 12-15-2015 06:55 PM | |
1185 | 11-06-2015 12:33 AM | |
1182 | 09-29-2015 10:11 PM |
02-11-2016
08:16 PM
1 Kudo
When you kill the topology you can specify a time that storm should wait before removing the topology completely. The default value for this wait time is 30 second. Can you verify that you are actually seeing the topology even after 30 seconds have elapsed or if you specify wait time explicitly , are you seeing the topology even after that time has elapsed?
... View more
01-14-2016
07:34 PM
1 Kudo
Please check your server.properties file and ensure you have authorizer.class.name set to Ranger Authorizer's Fully Qualified class name.
... View more
01-11-2016
09:22 PM
2 Kudos
No way to read based on timestamp right now, there is a discussion in the community https://cwiki.apache.org/confluence/display/KAFKA/KIP-32+-+Add+timestamps+to+Kafka+message
... View more
12-15-2015
06:55 PM
Storm does not support the storm-rabbitmq spout that you mentioning, it does not ship with apache storm distro so I have not looked at the code. Here is a JMS spout that you may be able to use as is with RabitMQ if RAbitMQ supports JMS https://github.com/hortonworks/storm/tree/2.3-mai... (Note that this is available in HWX repo but not under apache)
For an example of OpaqueSpout take a look at kafka-spout. https://github.com/apache/storm/blob/master/extern... The unit test part for exactly once is kind of hard in my opinion and there are no current examples. I am not sure if I understand the sharing queue's question completely. Your spout could read from N queues while outputting all the messages to a single stream, that is completely up to your spout. The merge can be useful if you want to create one spout -> one queue mapping and then merge all the output streams into a single stream that other processing units (trident states) subscribe to.
... View more
11-06-2015
12:33 AM
No way to kerberize Kafka in 2.2 If you are running storm-2.2 you should still be able to run a topology that uses storm-kafka connector's version from version 2.3 which should be able to read from a secure kafka cluster (HDP or not HDP).
... View more
09-29-2015
10:11 PM
You must specify "hdfs.storm.dir" in the topology config which should map to a hdfs directory where these files are stored. In addition hdfs-site.xml and core-site.xml must be part of nimbus's classpath. By default the hdfs jars are not shipped with storm so the user will have to set this up on all nimbus hosts and add these jars in nimbus's classpath using nimbus.childopts. In general it is better for now to use the default code distributor as HDFSCodeDistributor is written as an example and it is a class that will probably still evolve in near future.
... View more
09-29-2015
08:30 PM
What do you mean by multiple Kafka spouts? When you configure a spout to consume from a topic depending on what parallelism hint (ideally = number of kafka partitions) and what number of workers you set that single spout instance will be executed in different worker JVM's each one consuming from a kafka partition. You can find a trident example here https://github.com/apache/storm/blob/master/examples/storm-starter/src/jvm/storm/starter/trident/TridentKafkaWordCount.java
... View more