<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Not able to flume twitter data in to hdfs in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162724#M29318</link>
    <description>&lt;A rel="user" href="https://community.cloudera.com/users/6235/karthikgolagani.html" nodeid="6235"&gt;@karthik sai&lt;/A&gt;&lt;P&gt;Looks like you are using CDH distro therefore I would recommend you to run sam test on HDP cluster with Flume and let us know if you still face any issue.&lt;/P&gt;</description>
    <pubDate>Mon, 23 May 2016 20:09:48 GMT</pubDate>
    <dc:creator>jyadav</dc:creator>
    <dc:date>2016-05-23T20:09:48Z</dc:date>
    <item>
      <title>Not able to flume twitter data in to hdfs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162723#M29317</link>
      <description>&lt;P&gt;
	Hello team,&lt;/P&gt;&lt;P&gt;
	I'm a programming enthusiast.I have downloaded twitter stream before but now i'm not able to do so.I'm using apache-flume-1.4 on my hadoop 2.3.0 and cdh 5.0.0.&lt;/P&gt;&lt;P&gt;
	No matter how many times i've tried ,it is throwing the same error,&lt;/P&gt;&lt;PRE&gt;hadoop@ubuntu:~/hadoop/apache-flume-1.4.0-cdh5.0.0-bin$ ./bin/flume-ng agent -n TwitterAgent -c conf -f /home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/conf/local.conf Dflume.root.logger=DEBUG,console -n TwitterAgent


Info: Sourcing environment configuration script /home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/conf/flume-env.sh
Info: Including Hadoop libraries found via (/home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/bin/hadoop) for HDFS access
Info: Excluding /home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar from classpath
Info: Excluding /home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar from classpath
Info: Including HBASE libraries found via (/home/hadoop/hadoop/hbase-0.96.1.1-cdh5.0.0/bin/hbase) for HBASE access
Info: Excluding /home/hadoop/hadoop/hbase-0.96.1.1-cdh5.0.0/lib/slf4j-api-1.7.5.jar from classpath
Info: Excluding /home/hadoop/hadoop/hbase-0.96.1.1-cdh5.0.0/lib/slf4j-log4j12-1.7.5.jar from classpath
Info: Excluding /home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar from classpath
Info: Excluding /home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar from classpath
+ exec /usr/lib/jvm/java-7-openjdk-amd64/bin/java -Xms100m -Xmx200m -Dcom.sun.management.jmxremote -cp '/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/conf:/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/lib/*:/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/lib/flume-sources-1.0-SNAPSHOT.jar:/home/hadoop/hadoop/hadoop-2.3.0-cdh5.0.0/etc/hadoop:/home/ha.....

&lt;/PRE&gt;&lt;P&gt;And the .conf file is as follows:&lt;/P&gt;&lt;PRE&gt;TwitterAgent.sources= Twitter 
TwitterAgent.channels= MemChannel 
TwitterAgent.sinks=HDFS 
TwitterAgent.sources.Twitter.type = org.apache.flume.source.twitter.TwitterSource 
TwitterAgent.sources.Twitter.channels=MemChannel 
 
TwitterAgent.sources.Twitter.consumerKey=Pw63cpjptT59uT6w 
TwitterAgent.sources.Twitter.consumerSecret=    n8awrhKf7S576DcILPk5Ddfp1LQUU 
TwitterAgent.sources.Twitter.accessToken=163543326-s0Rqm5y4UC2WV7HPOuiOE9fPZZ56eWO95P 
TwitterAgent.sources.Twitter.accessTokenSecret=    CLwyJJ1jY4atf7iaiaR96Z1PmVvKF0iOXsP8E 
 
TwitterAgent.sources.Twitter.keywords= hadoop,election,sports, cricket,Big data,Trump 
 
TwitterAgent.sinks.HDFS.channel=MemChannel 
TwitterAgent.sinks.HDFS.type=hdfs 
TwitterAgent.sinks.HDFS.hdfs.path=hdfs://localhost:9000/tweety 
TwitterAgent.sinks.HDFS.hdfs.fileType=DataStream 
TwitterAgent.sinks.HDFS.hdfs.writeformat=Text 
TwitterAgent.sinks.HDFS.hdfs.batchSize=1000 
TwitterAgent.sinks.HDFS.hdfs.rollSize=0 
TwitterAgent.sinks.HDFS.hdfs.rollCount=10000 
TwitterAgent.sinks.HDFS.hdfs.rollInterval=600 
TwitterAgent.channels.MemChannel.type=memory 
TwitterAgent.channels.MemChannel.capacity=10000 
TwitterAgent.channels.MemChannel.transactionCapacity=100&lt;/PRE&gt;&lt;P&gt; And flume-env.sh file as follows:&lt;/P&gt;&lt;PRE&gt;# Enviroment variables can be set here.

JAVA_HOME=/usr/lib/jvm/java-7-openjdk-amd64

# Give Flume more memory and pre-allocate, enable remote monitoring via JMX
JAVA_OPTS="-Xms100m -Xmx200m -Dcom.sun.management.jmxremote"

# Note that the Flume conf directory is always included in the classpath.
FLUME_CLASSPATH="/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/lib/flume-sources-1.0-SNAPSHOT.jar"

&lt;/PRE&gt;&lt;P&gt; And the .bashrc file:&lt;/P&gt;&lt;PRE&gt;export FLUME_HOME="/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin"
export PATH="$FLUME_HOME/bin:$PATH"
export FLUME_CLASSPATH="$CLASSPATH:/home/hadoop/hadoop/apache-flume-1.4.0-cdh5.0.0-bin/lib/flume-sources-1.0-SNAPSHOT.jar "
&lt;/PRE&gt;&lt;P&gt;Please i want to know on which part i'm doing it wrong.&lt;/P&gt;&lt;P&gt;Any valuable suggestion is much appreciated.&lt;/P&gt;&lt;P&gt;Thanks in advance.&lt;/P&gt;</description>
      <pubDate>Mon, 23 May 2016 19:59:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162723#M29317</guid>
      <dc:creator>jack0188</dc:creator>
      <dc:date>2016-05-23T19:59:19Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to flume twitter data in to hdfs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162724#M29318</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/6235/karthikgolagani.html" nodeid="6235"&gt;@karthik sai&lt;/A&gt;&lt;P&gt;Looks like you are using CDH distro therefore I would recommend you to run sam test on HDP cluster with Flume and let us know if you still face any issue.&lt;/P&gt;</description>
      <pubDate>Mon, 23 May 2016 20:09:48 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162724#M29318</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-05-23T20:09:48Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to flume twitter data in to hdfs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162725#M29319</link>
      <description>&lt;P&gt;So, apache-flume-1.4 can still bring the data? or shall i upgrade my flume to 1.6 or higher?&lt;/P&gt;</description>
      <pubDate>Mon, 23 May 2016 20:09:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162725#M29319</guid>
      <dc:creator>jack0188</dc:creator>
      <dc:date>2016-05-23T20:09:49Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to flume twitter data in to hdfs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162726#M29320</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/6235/karthikgolagani.html" nodeid="6235"&gt;@karthik sai&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Hi Karthik, I was saying if you can install the Hortonworks Hadoop cluster or probably a Sandbox machine along with flume would help us to understand your issue while you run the same flume example on that.&lt;/P&gt;&lt;P&gt;Here is the download link of sandbox.&lt;/P&gt;&lt;P&gt;&lt;A href="http://hortonworks.com/downloads/#sandbox"&gt;http://hortonworks.com/downloads/#sandbox&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 23 May 2016 20:22:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162726#M29320</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-05-23T20:22:27Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to flume twitter data in to hdfs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162727#M29321</link>
      <description>&lt;P&gt;Able to flume twitter feeds in the sandbox after spending lot of time.&lt;/P&gt;&lt;P&gt;Following steps helped in resolving this:&lt;/P&gt;&lt;P&gt;1. Added below entry in /etc/hosts file&lt;/P&gt;&lt;P&gt;199.59.148.138 stream.twitter.com&lt;/P&gt;&lt;P&gt;2. updating datetime in sandbox&lt;/P&gt;&lt;P&gt;sudo ntpdate ntp.ubuntu.com&lt;/P&gt;&lt;P&gt;3. Adjusting hdfs path to point to 8020 port&lt;/P&gt;&lt;P&gt;TwitterAgent.sinks.HDFS.hdfs.path=hdfs://sandbox.hortonworks.com:8020/user/maria_dev/tweets/%Y/%m/%d/%H/&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jul 2017 07:27:44 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-flume-twitter-data-in-to-hdfs/m-p/162727#M29321</guid>
      <dc:creator>srinivask_kv</dc:creator>
      <dc:date>2017-07-11T07:27:44Z</dc:date>
    </item>
  </channel>
</rss>

