Member since
06-13-2016
30
Posts
20
Kudos Received
5
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
10549 | 12-26-2016 08:14 PM | |
4301 | 12-24-2016 05:22 PM | |
2977 | 12-19-2016 02:01 AM | |
5274 | 12-05-2016 01:52 PM | |
7412 | 09-08-2016 09:00 PM |
04-21-2017
12:50 PM
1 Kudo
@Mukesh Kumar Not sure why the article is protected, but it essentially says to switch the superset database to mysql/postgreSQL instead of sqlite for a temporary work around. You can adjust this in Ambari by going to the druid service and modifying the config for the "Superset Database name".
... View more
02-22-2017
02:19 PM
Is it possible to configure multiple notification emails for a SmartSense Account? Is there a delimiter that would allow adding more than just one email address, for instance: jim@domain.com; dan@domain.com?
... View more
Labels:
- Labels:
-
Hortonworks SmartSense
12-26-2016
08:14 PM
4 Kudos
@David Parkins Your best option for doing this is going to be using the
UpdateAttribute processor and then using a little NiFi Expression Language to
append the date onto your file. From
there you should be able to configure. I
created a sample flow and used the following config and achieved the following
results: /var/log/log1.log New filename: log1.log.2016-12-26 15:10:25.966Z This is the config on the updateAttribute processor: This is just a sample, you could do a lot better using append/prepend with NiFi Expression Language. There is more info here: https://nifi.apache.org/docs/nifi-docs/html/expression-language-guide.html
... View more
12-26-2016
06:29 PM
@mayki wogno
It should already be available in your script, if you use this:
flowFile = session.get()
flowFile.getAttribute('filename')
This should return your filename in your executeScript.
... View more
12-25-2016
03:21 PM
@Rohit Ravishankar That should be fine, you could also use splitText processor before the extractText processor. So basically each flowfile would contain a single record.
... View more
12-25-2016
01:59 PM
@Praveen PentaReddy Can you please mark an answer as accepted, and then create a new question with your new problem? This way it is indexed and other people with a similar new problem will be able to benefit? Please post your Flume conf file as well, i believe you might have a bind property off in it which is why the sink is not working. Thanks.
... View more
12-24-2016
05:22 PM
1 Kudo
@Praveen PentaReddyMost likely in your twitter.conf you have "TwitterAgent.sink.HDFS.channels" instead of "sinks". There will be multiple properties and they should all be TwitterAgent.sinks not "sink".
If that doesn't work you may want to post your twitter.conf and flume.env.sh so we can help narrow it down. I would also strongly consider looking at using NiFi for handling the movement of data from Twitter to HDFS, here is a detailed tutorial on doing that: https://community.hortonworks.com/articles/1282/sample-hdfnifi-flow-to-push-tweets-into-solrbanana.html
... View more
12-24-2016
04:58 PM
@qfdk Would you mind sending a copy of the data that comes out of the "replaceText" processor? I would like to try on my local system to post to ES. Do you have the mapping for that index as well in ES? Is the data as simple as: { "gid" : 799092, "coordinates" : [ 6.933880894889725, 48.0070784663541 ]},
{ "gid" : 799092, "coordinates" : [ 6.933880894889725, 48.0070784663541 ]},{ "gid" : 799092, "coordinates" : [ 6.933880894889725, 48.0070784663541 ]}
It looks to me like you should be using just the PutElasticSearch processor which uses the bulk api. I see in your "PostHTTP" processor in your video that you are pushing out to the "_bulk" endpoint. Looking at your error log, it seems it is not formatted correctly when being indexed in ES: "Caused by: java.lang.IllegalArgumentException: Malformed content, found extra data after parsing: START_OBJECT"
... View more
12-20-2016
06:28 PM
@qfdk In your flow, note that the "PutElasticSearch" processor actually uses the ES bulk api: https://www.elastic.co/guide/en/elasticsearch/client/java-api/current/java-docs-bulk.html. So if you are specifying a reasonable amount in the batch size you should essentially be "bulk inserting". What exception in index did you receive from ElasticSearch? Were you pushing documents/data to the wrong index with an incorrect mapping? Posting the error from ES will help. The bulk index api in ES can sometimes be a bit clunky, so instead I use the ES HTTP processor and adjust the batchsize accordingly. This makes sending the raw JSON to the ES a little easier and straightforward. For the data being stuck, can you run a nifi.sh dump and then post the contents of the bootstrap.log?
... View more
12-19-2016
02:01 AM
@Kumar The NiFi provenance UI adds a newline to your content. If you go look at the raw file being fetched or moved through the FetchFile processor you will notice that there actually isn't a new line appended. If I use the same input you do, my provenance UI will show a new line (line 2) after the text "Line1", identical to your screenshot. However if i look at where the "move" content of the fetchFile Processor is placed, i see that there isn't a newline. Also if i do a putFile i notice there isn't a newline either. Below are some screenshots to show: Raw file: Provenance shows a new line (line 2): The output of Putfile: The output of having MoveFile in the fetchFile processor: The overall flow used to test:
... View more