<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Flume Tutorials in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128156#M90853</link>
    <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/140/nsabharwal.html" nodeid="140"&gt;@Neeraj Sabharwal&lt;/A&gt; Thanks!  I hadn't ran across this one yet.&lt;/P&gt;</description>
    <pubDate>Fri, 05 Feb 2016 21:41:36 GMT</pubDate>
    <dc:creator>daniel_hendrix</dc:creator>
    <dc:date>2016-02-05T21:41:36Z</dc:date>
    <item>
      <title>Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128152#M90849</link>
      <description>&lt;P&gt;Are there any good walkthrough tutorials for Flume?  I've seen the two listed &lt;A href="http://hortonworks.com/hadoop/flume/#tutorials"&gt;here&lt;/A&gt;.  However, after skimming through the second one "Analyzing Social Media and Customer Sentiment," I fail to see any use or reference of Flume within it.  I would specifically like something that walks through performance of the two Flume objectives documented in the HDP Certified Developer Exam Objectives sheet:&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;Given a Flume configuration file, start a Flume agent&lt;/LI&gt;&lt;LI&gt;Given a configured sink and source, configure a Flume memory channel with a specified capacity  https
&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;The 1st tutorial from the link above starts a Flume agent via Ambari, but I assume the Exam will require this to be done via the Terminal.  &lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:27:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128152#M90849</guid>
      <dc:creator>daniel_hendrix</dc:creator>
      <dc:date>2016-02-05T21:27:10Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128153#M90850</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/2096/danielhendrix.html" nodeid="2096"&gt;@Daniel Hendrix&lt;/A&gt;
&lt;/P&gt;&lt;P&gt;This is good tutorial &lt;A href="http://hortonworks.com/blog/configure-elastic-search-hadoop-hdp-2-0/" target="_blank"&gt;http://hortonworks.com/blog/configure-elastic-search-hadoop-hdp-2-0/&lt;/A&gt;  "flume example"&lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:28:58 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128153#M90850</guid>
      <dc:creator>nsabharwal</dc:creator>
      <dc:date>2016-02-05T21:28:58Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128154#M90851</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/164/rich.html" nodeid="164"&gt;@rich&lt;/A&gt;        &lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:29:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128154#M90851</guid>
      <dc:creator>nsabharwal</dc:creator>
      <dc:date>2016-02-05T21:29:29Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128155#M90852</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/2096/danielhendrix.html" nodeid="2096"&gt;@Daniel Hendrix&lt;/A&gt;&lt;P&gt;I wrote a small tutorial on flume as a how-to rather than use-case based as the ones you specified. I think by far the best resource would be the Flume &lt;A href="http://flume.apache.org/"&gt;website&lt;/A&gt;. It has examples for every possible sink, source and channel. Please see below&lt;/P&gt;&lt;PRE&gt;Flume

# HDP 2.3.2 Sandbox
# Example, single-node Flume configuration using netcat source, memory channel and logger sink

# install telnet
yum install -y telnet

# start flume with this configuration
******************************************************************************
# example.conf: A single-node Flume configuration
# Name the components on this agent
a1.sources = r1
a1.sinks = k1
a1.channels = c1
# Describe/configure the source
a1.sources.r1.type = netcat
a1.sources.r1.bind = localhost
a1.sources.r1.port = 44444
# Describe the sink
a1.sinks.k1.type = logger
# Use a channel which buffers events in memory
a1.channels.c1.type = memory
a1.channels.c1.capacity = 1000
a1.channels.c1.transactionCapacity = 100
# Bind the source and sink to the channel
a1.sources.r1.channels = c1
a1.sinks.k1.channel = c1
******************************************************************************
# in another terminal
telnet localhost 44444

# type anything
# then in the original terminal
tail -f /var/log/flume/flume-a1.log

# Exampe netcat source, hdfs sink as DataStream
# create hdfs flume directory
sudo -u hdfs hdfs dfs -mkdir /flume
sudo -u hdfs hdfs dfs -mkdir /flume/events
sudo -u hdfs hdfs dfs -chown -R flume:hdfs /flume/events

******************************************************************************
# example.conf: A single-node Flume configuration
# Name the components on this agent
a1.sources = r1
a1.sinks = k1
a1.channels = c1
# Describe/configure the source
a1.sources.r1.type = netcat
a1.sources.r1.bind = localhost
a1.sources.r1.port = 44444
# Describe the sink
 
a1.sinks.k1.type = hdfs
a1.sinks.k1.hdfs.path = /flume/events/%y-%m-%d/%H%M/%S
a1.sinks.k1.hdfs.filePrefix = events-
a1.sinks.k1.hdfs.round = true
a1.sinks.k1.hdfs.roundValue = 10
a1.sinks.k1.hdfs.roundUnit = minute
a1.sinks.k1.hdfs.useLocalTimeStamp = true
a1.sinks.k1.hdfs.fileType = DataStream

# Use a channel which buffers events in memory
a1.channels.c1.type = memory
a1.channels.c1.capacity = 1000
a1.channels.c1.transactionCapacity = 100
# Bind the source and sink to the channel
a1.sources.r1.channels = c1
a1.sinks.k1.channel = c1
******************************************************************************

# show the output in hdfs
sudo -u flume hdfs dfs -ls /flume/events/
sudo -u flume hdfs dfs -ls /flume/events/
sudo -u flume hdfs dfs -cat /flume/events/*/*/*/*&lt;/PRE&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:31:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128155#M90852</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-05T21:31:16Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128156#M90853</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/140/nsabharwal.html" nodeid="140"&gt;@Neeraj Sabharwal&lt;/A&gt; Thanks!  I hadn't ran across this one yet.&lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:41:36 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128156#M90853</guid>
      <dc:creator>daniel_hendrix</dc:creator>
      <dc:date>2016-02-05T21:41:36Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128157#M90854</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/393/aervits.html" nodeid="393"&gt;@Artem Ervits&lt;/A&gt; Thanks, this is very helpful.&lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 21:42:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128157#M90854</guid>
      <dc:creator>daniel_hendrix</dc:creator>
      <dc:date>2016-02-05T21:42:28Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128158#M90855</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/2096/danielhendrix.html" nodeid="2096"&gt;@Daniel Hendrix&lt;/A&gt;&lt;/P&gt;&lt;P&gt;You are correct - you should be able to start a Flume agent from the command line. The docs show how to do this:&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.cloudera.com/"&gt;https://flume.apache.org/FlumeUserGuide.html#starting-an-agent&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.cloudera.com/"&gt;&lt;/A&gt;You also need to know how to configure a memory channel, which is also demonstrated in the docs:&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.cloudera.com/"&gt;https://flume.apache.org/FlumeUserGuide.html#memory-channel&lt;/A&gt;&lt;/P&gt;&lt;P&gt;I would recommend going through the same tutorial that you found here:&lt;/P&gt;&lt;P&gt; &lt;A href="https://community.cloudera.com/"&gt;http://hortonworks.com/hadoop-tutorial/how-to-refine-and-visualize-server-log-data/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Within that tutorial, configure a memory channel and try starting it from the command line. Let me know if you have any issues along the way and I'll be glad to assist.&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;Rich Raposa&lt;/P&gt;&lt;P&gt;Certification Manager&lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 22:16:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128158#M90855</guid>
      <dc:creator>rich1</dc:creator>
      <dc:date>2016-02-05T22:16:49Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128159#M90856</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/164/rich.html" nodeid="164"&gt;@rich&lt;/A&gt;  This is perfect for an article..Thanks , Rich! &lt;A rel="user" href="https://community.cloudera.com/users/2096/danielhendrix.html" nodeid="2096"&gt;@Daniel Hendrix&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 05 Feb 2016 22:52:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128159#M90856</guid>
      <dc:creator>nsabharwal</dc:creator>
      <dc:date>2016-02-05T22:52:09Z</dc:date>
    </item>
    <item>
      <title>Re: Flume Tutorials</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128160#M90857</link>
      <description>&lt;P&gt;Hi Daniel,&lt;/P&gt;&lt;P&gt;Were you able to run this flume example?
Because I am trying.

What would be the values if the sink is hdfs and not elasticsearch? any idea&lt;/P&gt;&lt;P&gt;Thank you.&lt;/P&gt;</description>
      <pubDate>Sun, 10 Apr 2016 22:18:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Flume-Tutorials/m-p/128160#M90857</guid>
      <dc:creator>Eukrev</dc:creator>
      <dc:date>2016-04-10T22:18:03Z</dc:date>
    </item>
  </channel>
</rss>

