Member since
07-08-2013
26
Posts
8
Kudos Received
3
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
14796 | 01-09-2015 07:09 AM | |
10063 | 12-29-2014 09:57 AM | |
3617 | 12-23-2014 05:55 PM |
01-15-2015
01:57 PM
Thanks Joey, looks like "/usr/bin/flume-ng agent ... -Duser.home=/user/xyz -c /path/to/jar1:/path/to/jar2" seem to have done the trick by prepending to the java.class.path.
... View more
01-09-2015
07:28 AM
Keep in mind that with the MemoryChannel you lose any records in the channel if Flume crashes or the system reboots.
... View more
12-29-2014
09:57 AM
1 Kudo
If you want each file to end up remaining whole, you can use the BlobDeserialzier[1] for the deserializer parameter of the SpoolingDirectorySource[2].: a1.channels = c1
a1.sources = src-1
a1.sources.src-1.type = spooldir
a1.sources.src-1.channels = ch-1
a1.sources.src-1.spoolDir = /var/log/apache/flumeSpool
a1.sources.src-1.fileHeader = true a1.sources.src-1.deserializer = org.apache.flume.sink.solr.morphline.BlobDeserializer$Builder If you need to, set deserialzier.maxBlobLength to the maximum file size you'll be picking up. The default is 100 million bytes. This won't work for very large files as the entire file contents will get buffered into RAM. The File channel is the best option for reliable data flow. If you want the output file to have the same name is the input file, you can set the basenameHeader parameter to true. This will set a header in the flume event called basename. You can customize the name of the header by setting basenameHeaderKey. Then in your sink configuration, you can refer to the header value in the filePrefix with something like this: a1.channels = c1
a1.sinks = k1
a1.sinks.k1.type = hdfs
a1.sinks.k1.channel = c1
a1.sinks.k1.hdfs.path = /flume/events/
a1.sinks.k1.hdfs.filePrefix = %{basename}- a1.sinks.k1.hdfs.fileType = DataStream HTH, -Joey [1] http://flume.apache.org/FlumeUserGuide.html#blobdeserializer [2] http://flume.apache.org/FlumeUserGuide.html#spooling-directory-source
... View more
12-24-2014
02:57 PM
I had to delete the directories in HDFS manually It could be that kite-dataset delete command only does a logical delete. this means it only removes the metadata. anyways. doing kite-dataset delete and then a manual delete in HDFS works for me.
... View more