<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Unable to deliver event. Exception follows in flume in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-deliver-event-Exception-follows-in-flume/m-p/39312#M24268</link>
    <description>&lt;P&gt;Hi, I got it by changing roll size and batch size now its working fine.&lt;/P&gt;&lt;P&gt;rollSize = 100000 and batchsize=100&lt;/P&gt;</description>
    <pubDate>Mon, 04 Apr 2016 10:30:53 GMT</pubDate>
    <dc:creator>Tejaponnaluru</dc:creator>
    <dc:date>2016-04-04T10:30:53Z</dc:date>
    <item>
      <title>Unable to deliver event. Exception follows in flume</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-deliver-event-Exception-follows-in-flume/m-p/39306#M24267</link>
      <description>&lt;P&gt;I'm trying to load data from local to hdfs using spooldir source getting process failed error&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here My error&lt;/P&gt;&lt;PRE&gt;process failed
org.apache.flume.ChannelException: Take list for MemoryTransaction, capacity 100 full, consider committing more frequently, increasing capacity, or increasing thread count
	at org.apache.flume.channel.MemoryChannel$MemoryTransaction.doTake(MemoryChannel.java:96)
	at org.apache.flume.channel.BasicTransactionSemantics.take(BasicTransactionSemantics.java:113)
	at org.apache.flume.channel.BasicChannelSemantics.take(BasicChannelSemantics.java:95)
	at org.apache.flume.sink.hdfs.HDFSEventSink.process(HDFSEventSink.java:374)
	at org.apache.flume.sink.DefaultSinkProcessor.process(DefaultSinkProcessor.java:68)
	at org.apache.flume.SinkRunner$PollingRunner.run(SinkRunner.java:147)
	at java.lang.Thread.run(Thread.java:745)&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;and here my confi file&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Local agent&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;agent.sources.localsource.type = spooldir&lt;BR /&gt;#agent.sources.localsource.shell = /bin/bash -c&lt;BR /&gt;agent.sources.localsource.spoolDir = /home/dwh/teja/Flumedata/&lt;BR /&gt;agent.sources.localsource.fileHeader = true&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# The channel can be defined as follows.&lt;BR /&gt;agent.sources.localsource.channels = memoryChannel&lt;/P&gt;&lt;P&gt;# Each sink's type must be defined&lt;BR /&gt;agent.sinks.avro_Sink.type = avro&lt;BR /&gt;agent.sinks.avro_Sink.hostname=192.168.4.444&lt;BR /&gt;agent.sinks.avro_Sink.port= 8021&lt;BR /&gt;agent.sinks.avro_Sink.avro.batchSize = 1000&lt;BR /&gt;agent.sinks.avro_Sink.avro.rollCount = 0&lt;BR /&gt;agent.sinks.avro_Sink.avro.rollSize = 1000000&lt;BR /&gt;agent.sinks.avro_Sink.avro.rollInterval = 0&lt;BR /&gt;agent.sinks.avro_Sink.channel = memoryChannel&lt;/P&gt;&lt;P&gt;# Each channel's type is defined.&lt;BR /&gt;agent.channels.memoryChannel.type = memory&lt;BR /&gt;# In this case, it specifies the capacity of the memory channel&lt;BR /&gt;agent.channels.memoryChannel.capacity = 10000&lt;BR /&gt;agent.channels.memoryChannel.transactionCapacity = 10000&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Remote config file&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;# Please paste flume.conf here. Example:&lt;/P&gt;&lt;P&gt;# Sources, channels, and sinks are defined per&lt;BR /&gt;# agent name, in this case 'tier1'.&lt;BR /&gt;tier1.sources = source1&lt;BR /&gt;tier1.channels = channel1&lt;BR /&gt;tier1.sinks = sink1&lt;BR /&gt;tier1.sources.source1.type = avro&lt;BR /&gt;tier1.sources.source1.bind = 192.168.4.444&lt;BR /&gt;tier1.sources.source1.port=8021&lt;BR /&gt;tier1.sources.source1.channels = channel1&lt;BR /&gt;tier1.channels.channel1.type = memory&lt;BR /&gt;tier1.sinks.sink1.type = hdfs&lt;BR /&gt;tier1.sinks.sink1.channel = channel1&lt;BR /&gt;tier1.sinks.sink1.hdfs.path = hdfs://192.168.4.444:8020/user/hadoop/flumelogs/&lt;BR /&gt;tier1.sinks.sink1.hdfs.fileType = DataStream&lt;BR /&gt;tier1.sinks.sink1.hdfs.writeFormat= Text&lt;BR /&gt;tier1.sinks.sink1.hdfs.batchSize = 1000&lt;BR /&gt;tier1.sinks.sink1.hdfs.rollCount = 0&lt;BR /&gt;tier1.sinks.sink1.hdfs.rollSize = 1000000&lt;BR /&gt;tier1.sinks.sink1.hdfs.rollInterval = 0&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;tier1.channels.channel1.capacity = 10000&lt;BR /&gt;tier1.channels.channel1.transactioncapacity=10000&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please Help.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:12:08 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-deliver-event-Exception-follows-in-flume/m-p/39306#M24267</guid>
      <dc:creator>Tejaponnaluru</dc:creator>
      <dc:date>2022-09-16T10:12:08Z</dc:date>
    </item>
    <item>
      <title>Re: Unable to deliver event. Exception follows in flume</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-deliver-event-Exception-follows-in-flume/m-p/39312#M24268</link>
      <description>&lt;P&gt;Hi, I got it by changing roll size and batch size now its working fine.&lt;/P&gt;&lt;P&gt;rollSize = 100000 and batchsize=100&lt;/P&gt;</description>
      <pubDate>Mon, 04 Apr 2016 10:30:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-deliver-event-Exception-follows-in-flume/m-p/39312#M24268</guid>
      <dc:creator>Tejaponnaluru</dc:creator>
      <dc:date>2016-04-04T10:30:53Z</dc:date>
    </item>
  </channel>
</rss>

