<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Import  Data from MongoDb to HBase using Spark in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148351#M40446</link>
    <description>&lt;P&gt;did you try this?&lt;/P&gt;&lt;P&gt;  &lt;A href="https://github.com/colinmarc/zerowing" target="_blank"&gt;https://github.com/colinmarc/zerowing&lt;/A&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 13 Sep 2016 13:02:09 GMT</pubDate>
    <dc:creator>rajkumar_singh</dc:creator>
    <dc:date>2016-09-13T13:02:09Z</dc:date>
    <item>
      <title>Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148350#M40445</link>
      <description>&lt;P&gt;Hi All,&lt;/P&gt;&lt;P&gt;Can someone please kind to tell me how to import the data from MongoDB to HBase using Spark or Without Using Spark.&lt;/P&gt;&lt;P&gt;If not any other Way.&lt;/P&gt;&lt;P&gt;Regards,&lt;/P&gt;&lt;P&gt;Vijay&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 12:55:13 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148350#M40445</guid>
      <dc:creator>vijaykumar243</dc:creator>
      <dc:date>2016-09-13T12:55:13Z</dc:date>
    </item>
    <item>
      <title>Re: Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148351#M40446</link>
      <description>&lt;P&gt;did you try this?&lt;/P&gt;&lt;P&gt;  &lt;A href="https://github.com/colinmarc/zerowing" target="_blank"&gt;https://github.com/colinmarc/zerowing&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 13:02:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148351#M40446</guid>
      <dc:creator>rajkumar_singh</dc:creator>
      <dc:date>2016-09-13T13:02:09Z</dc:date>
    </item>
    <item>
      <title>Re: Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148352#M40447</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/8919/rajkumarsingh.html" nodeid="8919"&gt;@Rajkumar Singh&lt;/A&gt;&lt;/P&gt;&lt;P&gt;I did tried this example. But my Mongo Json data is Very complex&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 14:04:15 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148352#M40447</guid>
      <dc:creator>vijaykumar243</dc:creator>
      <dc:date>2016-09-13T14:04:15Z</dc:date>
    </item>
    <item>
      <title>Re: Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148353#M40448</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/11346/vijaykumar243.html" nodeid="11346"&gt;@Vijay Kumar J&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Have you considered using Apache NiFi for this?&lt;/P&gt;&lt;P&gt;NiFi has inbuilt processors to work with data in both MongoDB and HBase.&lt;/P&gt;&lt;P&gt;You could use NiFi's &lt;STRONG&gt;GetMongo&lt;/STRONG&gt; processor followed by the &lt;STRONG&gt;PutHbaseJSON &lt;/STRONG&gt;processor to move the data from MongoDB to HBase.&lt;/P&gt;&lt;P&gt;Check out the following article for more info on using NiFi to interact with MongoDB:&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.hortonworks.com/articles/53554/using-apache-nifi-100-with-mongodb.html"&gt;https://community.hortonworks.com/articles/53554/using-apache-nifi-100-with-mongodb.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 14:37:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148353#M40448</guid>
      <dc:creator>ldaluz</dc:creator>
      <dc:date>2016-09-13T14:37:03Z</dc:date>
    </item>
    <item>
      <title>Re: Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148354#M40449</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/94/ldaluz.html" nodeid="94"&gt;@Laurence Da Luz&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Thanks for the response.&lt;/P&gt;&lt;P&gt;What if the process fails while fetching (or) storing the data.&lt;/P&gt;&lt;P&gt;Is it feasible for production use.&lt;/P&gt;&lt;P&gt;How to execute this NIFI on Hourly basis.&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 17:47:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148354#M40449</guid>
      <dc:creator>vijaykumar243</dc:creator>
      <dc:date>2016-09-13T17:47:16Z</dc:date>
    </item>
    <item>
      <title>Re: Import  Data from MongoDb to HBase using Spark</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148355#M40450</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/11346/vijaykumar243.html" nodeid="11346"&gt;@Vijay Kumar J&lt;/A&gt;&lt;P&gt;NiFi is definitely feasible for production use, and it is perfectly suited for your MongoDB to HBase data movement use case. NiFi is a tool used for managing dataflow and integration between systems in an automated and configurable way. It allows you to stream, transform, and sort data and uses a drag-and-drop UI.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Dealing with failures - &lt;/STRONG&gt;NiFi is configurable - when you build your flow within NiFi you can determine how you want to handle failures. In your case, you could build a flow in NiFi that retries on failure, and sends out an email on failure (this is an example, how you want to handle failures for fetch and storing data can be configured however you need)&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Execute NiFi on an Hourly Basis &lt;/STRONG&gt;- NiFi isn't like traditional data movement schedulers, and flows built using NiFi are treated as 'always-on' where data can be constantly streamed as it is received. That being said, NiFi provides the ability to schedule each processor if needed, so in your case you could have your GetMongo processor set to run once every hour, and your PutHBaseJSON processor to push data to HBase as soon as it is received from the GetMongo processor.&lt;/P&gt;&lt;P&gt;Check out this tutorial for getting started and building your first flow: &lt;A href="http://hortonworks.com/hadoop-tutorial/learning-ropes-apache-nifi/"&gt;http://hortonworks.com/hadoop-tutorial/learning-ropes-apache-nifi/&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 13 Sep 2016 19:46:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Import-Data-from-MongoDb-to-HBase-using-Spark/m-p/148355#M40450</guid>
      <dc:creator>ldaluz</dc:creator>
      <dc:date>2016-09-13T19:46:03Z</dc:date>
    </item>
  </channel>
</rss>

