<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Flume--ng command error while importing weblog from webserver ip address to hadoop hdfs path in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42808#M34487</link>
    <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;I am using cloudera VM with CDH5.5.0.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;I am trying to pull weblog data using flume from /var/log/wtmp at ip address 10.3.9.34 at port 22.Let me inform i did ssh root@10.3.9.34 from command prompt of CDH5.5 and i was able to connect to this weblog ipaddress&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am trying to pull weblog from this ipaddress and put that weblog into hdfs path /user/cloudera/flume/ so i ran below flume-ng command :-&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;flume-ng agent --name agent1 --conf /home/cloudera/flume/conf --conf -file /home/cloudera/flume/conf/flume.conf&lt;/P&gt;&lt;P&gt;Problem is i am getting Fatal error as "java.lang.NullPointerException" while Import&lt;/P&gt;&lt;P&gt;Below is my flume.conf details :-&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;agent1.sources = netcat-collect&lt;BR /&gt;agent1.sinks = hdfs-write&lt;BR /&gt;agent1.channels = memory&lt;/P&gt;&lt;P&gt;# Describe/configure source1&lt;BR /&gt;agent1.sources.netcat-collect.type = netcat&lt;BR /&gt;agent1.sources.netcat-collect.bind = 10.3.9.34&lt;BR /&gt;agent1.sources.netcat-collect.port = 22&lt;BR /&gt;agent1.sources.netcat-collect.command = tail -F /var/log/wtmp&lt;/P&gt;&lt;P&gt;# Describe solrSink&lt;BR /&gt;agent1.sinks.hdfs-write.type = hdfs&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.path = /user/cloudera/flume/%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.filePrefix = flume-%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollSize = 1048576&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollCount = 100&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollInterval = 120&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.writeFormat = Text&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.fileType = DataStream&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.useLocalTimeStamp = true&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.idleTimeout = 10&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# Use a channel which buffers events to a file&lt;BR /&gt;# -- The component type name, needs to be FILE.&lt;BR /&gt;agent1.channels.memoryChannel.type = memory&lt;BR /&gt;agent1.channels.memoryChannel.capacity =10000&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# Amount of time (in millis) between checkpoints&lt;BR /&gt;agent1.channels.memoryChannel.checkpointInterval 3000&lt;/P&gt;&lt;P&gt;# Max size (in bytes) of a single log file&lt;BR /&gt;agent1.channels.memoryChannel.maxFileSize = 2146435071&lt;/P&gt;&lt;P&gt;# Bind the source and sink to the channel&lt;BR /&gt;agent1.sources.netcat-collect.channels = memoryChannel&lt;BR /&gt;agent1.sinks.hdfs-write.channel = memoryChannel&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Execution log attached with this thread&lt;/P&gt;&lt;P&gt;&lt;A title="Flume--ng execution log" href="https://drive.google.com/file/d/0B7FLyvHGgEJaYnM2d3JfRXMwNEU/view?usp=sharing" target="_blank"&gt;https://drive.google.com/file/d/0B7FLyvHGgEJaYnM2d3JfRXMwNEU/view?usp=sharing&lt;/A&gt;&lt;BR /&gt;Can someone help me in guiding what is the resolution&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 10:29:34 GMT</pubDate>
    <dc:creator>Tdas</dc:creator>
    <dc:date>2022-09-16T10:29:34Z</dc:date>
    <item>
      <title>Flume--ng command error while importing weblog from webserver ip address to hadoop hdfs path</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42808#M34487</link>
      <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;I am using cloudera VM with CDH5.5.0.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;I am trying to pull weblog data using flume from /var/log/wtmp at ip address 10.3.9.34 at port 22.Let me inform i did ssh root@10.3.9.34 from command prompt of CDH5.5 and i was able to connect to this weblog ipaddress&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am trying to pull weblog from this ipaddress and put that weblog into hdfs path /user/cloudera/flume/ so i ran below flume-ng command :-&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;flume-ng agent --name agent1 --conf /home/cloudera/flume/conf --conf -file /home/cloudera/flume/conf/flume.conf&lt;/P&gt;&lt;P&gt;Problem is i am getting Fatal error as "java.lang.NullPointerException" while Import&lt;/P&gt;&lt;P&gt;Below is my flume.conf details :-&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;agent1.sources = netcat-collect&lt;BR /&gt;agent1.sinks = hdfs-write&lt;BR /&gt;agent1.channels = memory&lt;/P&gt;&lt;P&gt;# Describe/configure source1&lt;BR /&gt;agent1.sources.netcat-collect.type = netcat&lt;BR /&gt;agent1.sources.netcat-collect.bind = 10.3.9.34&lt;BR /&gt;agent1.sources.netcat-collect.port = 22&lt;BR /&gt;agent1.sources.netcat-collect.command = tail -F /var/log/wtmp&lt;/P&gt;&lt;P&gt;# Describe solrSink&lt;BR /&gt;agent1.sinks.hdfs-write.type = hdfs&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.path = /user/cloudera/flume/%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.filePrefix = flume-%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollSize = 1048576&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollCount = 100&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollInterval = 120&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.writeFormat = Text&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.fileType = DataStream&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.useLocalTimeStamp = true&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.idleTimeout = 10&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# Use a channel which buffers events to a file&lt;BR /&gt;# -- The component type name, needs to be FILE.&lt;BR /&gt;agent1.channels.memoryChannel.type = memory&lt;BR /&gt;agent1.channels.memoryChannel.capacity =10000&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# Amount of time (in millis) between checkpoints&lt;BR /&gt;agent1.channels.memoryChannel.checkpointInterval 3000&lt;/P&gt;&lt;P&gt;# Max size (in bytes) of a single log file&lt;BR /&gt;agent1.channels.memoryChannel.maxFileSize = 2146435071&lt;/P&gt;&lt;P&gt;# Bind the source and sink to the channel&lt;BR /&gt;agent1.sources.netcat-collect.channels = memoryChannel&lt;BR /&gt;agent1.sinks.hdfs-write.channel = memoryChannel&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Execution log attached with this thread&lt;/P&gt;&lt;P&gt;&lt;A title="Flume--ng execution log" href="https://drive.google.com/file/d/0B7FLyvHGgEJaYnM2d3JfRXMwNEU/view?usp=sharing" target="_blank"&gt;https://drive.google.com/file/d/0B7FLyvHGgEJaYnM2d3JfRXMwNEU/view?usp=sharing&lt;/A&gt;&lt;BR /&gt;Can someone help me in guiding what is the resolution&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:29:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42808#M34487</guid>
      <dc:creator>Tdas</dc:creator>
      <dc:date>2022-09-16T10:29:34Z</dc:date>
    </item>
    <item>
      <title>Re: Flume--ng command error while importing weblog from webserver ip address to hadoop hdfs path</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42865#M34488</link>
      <description>&lt;P&gt;Thank you this got solved by below configuration and command&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;agent1.sources = netcat-collect&lt;BR /&gt;agent1.sinks = hdfs-write&lt;BR /&gt;agent1.channels = memoryChannel&lt;/P&gt;&lt;P&gt;# Describe/configure source1&lt;BR /&gt;agent1.sources.netcat-collect.type = exec&lt;BR /&gt;agent1.sources.netcat-collect.bind = 10.3.9.34&lt;BR /&gt;agent1.sources.netcat-collect.port = 22&lt;BR /&gt;agent1.sources.netcat-collect.command = tail -F /var/log/wtmp&lt;/P&gt;&lt;P&gt;# Describe solrSink&lt;BR /&gt;agent1.sinks.hdfs-write.type = hdfs&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.path = /user/cloudera/flume/%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.filePrefix = flume-%y-%m-%d&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollSize = 1048576&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollCount = 100&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.rollInterval = 120&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.writeFormat = Text&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.fileType = DataStream&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.useLocalTimeStamp = true&lt;BR /&gt;agent1.sinks.hdfs-write.hdfs.idleTimeout = 10&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;# Use a channel which buffers events to a file&lt;BR /&gt;# -- The component type name, needs to be FILE.&lt;BR /&gt;agent1.channels.memoryChannel.type = memory&lt;BR /&gt;agent1.channels.memoryChannel.capacity =10000&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;# Amount of time (in millis) between checkpoints&lt;BR /&gt;agent1.channels.memoryChannel.checkpointInterval 300000&lt;/P&gt;&lt;P&gt;# Max size (in bytes) of a single log file&lt;BR /&gt;agent1.channels.memoryChannel.maxFileSize = 2146435071&lt;/P&gt;&lt;P&gt;# Bind the source and sink to the channel&lt;BR /&gt;agent1.sources.netcat-collect.channels = memoryChannel&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Below is the command to pull data from weblog to HDFS&lt;/P&gt;&lt;P&gt;flume-ng agent --name agent1 --conf /home/cloudera/flume/conf --conf-file /home/cloudera/flume/conf/flume.conf&lt;/P&gt;</description>
      <pubDate>Thu, 14 Jul 2016 19:57:47 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42865#M34488</guid>
      <dc:creator>Tdas</dc:creator>
      <dc:date>2016-07-14T19:57:47Z</dc:date>
    </item>
    <item>
      <title>Re: Flume--ng command error while importing weblog from webserver ip address to hadoop hdfs path</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42870#M34489</link>
      <description>If you are using the exec source to tail a file, keep in mind that it is not a very reliable source. I would suggest using the taildir source (&lt;A href="https://archive.cloudera.com/cdh5/cdh/5/flume-ng/FlumeUserGuide.html#taildir-source" target="_blank"&gt;https://archive.cloudera.com/cdh5/cdh/5/flume-ng/FlumeUserGuide.html#taildir-source&lt;/A&gt;) to tail files reliably.&lt;BR /&gt;&lt;BR /&gt;-pd</description>
      <pubDate>Thu, 14 Jul 2016 21:02:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Flume-ng-command-error-while-importing-weblog-from-webserver/m-p/42870#M34489</guid>
      <dc:creator>pdvorak</dc:creator>
      <dc:date>2016-07-14T21:02:10Z</dc:date>
    </item>
  </channel>
</rss>

