<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Tracking of Hive tables metadata changes in real time using Atlas in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193378#M155442</link>
    <description>&lt;P&gt;&lt;STRONG&gt;@&lt;/STRONG&gt;&lt;A href="https://community.hortonworks.com/users/10331/mrizvi.html"&gt;Mushtaq Rizvi&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Yes. Please follow the instructions on how to add HDF components to an existent HDP 2.6.1 cluster:&lt;/P&gt;&lt;P&gt;&lt;A href="https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.0.1/bk_installing-hdf-on-hdp/content/upgrading_ambari.html" target="_blank"&gt;https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.0.1/bk_installing-hdf-on-hdp/content/upgrading_ambari.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;This is not the latest HDF, but it is compatible with HDP 2.6.1 and I was pretty happy with its stability and recommend it.&lt;/P&gt;&lt;P&gt; You would be able to add Apache NiFi 1.5, but also Schema Registry. NiFi Registry is part of the latest HDF 3.1.x, however, you would have to install it in a separate cluster and it is not worth it the effort for what you are trying to achieve right now. I would proceed with HDP upgrade when you are ready for HDF 3.2 which will be probably launched in the next couple months. &lt;/P&gt;&lt;P&gt;In case that you can't add another node to your cluster for NiFi, try to use one of the nodes that has low CPU utilization and some disk available for NiFi lineage data storage. It depends on how much lineage you want to preserve, but you should be probably fine with several tens of GB for starters.&lt;/P&gt;&lt;P&gt;If this response helped, please vote and accept answer.&lt;/P&gt;</description>
    <pubDate>Wed, 28 Mar 2018 11:22:52 GMT</pubDate>
    <dc:creator>cstanca</dc:creator>
    <dc:date>2018-03-28T11:22:52Z</dc:date>
    <item>
      <title>Tracking of Hive tables metadata changes in real time using Atlas</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193375#M155439</link>
      <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;
I am using HDP 2.6 and I want to track Hive tables metadata changes in real time. I have HiveHook enabled and I can see Kafka JSON messages in ATLAS_HOOK and ATLAS_ENTITIES topics. 
Also, Atlas is able to consume these entity updates.  I am looking for the most optimal way to get entity updates info real time. &lt;/P&gt;&lt;P&gt;1) Is there a way to create a NotificationServer (like SMTP) to which Atlas will send these updates? &lt;/P&gt;&lt;P&gt;2) Or do I have to create a custom Kafka consumer that reads data directly from ATLAS_HOOK or ATLAS_ENTITIES topics in JSON? &lt;/P&gt;&lt;P&gt;
P.S - I do not want to read everything from Kafka topic. There are thousands of tables but I want metadata changes for specific tables only. 
Please let me know how to setup the offsets related to particular databases/tables only. Thanks&lt;/P&gt;</description>
      <pubDate>Tue, 27 Mar 2018 02:12:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193375#M155439</guid>
      <dc:creator>mrizvi</dc:creator>
      <dc:date>2018-03-27T02:12:39Z</dc:date>
    </item>
    <item>
      <title>Re: Tracking of Hive tables metadata changes in real time using Atlas</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193376#M155440</link>
      <description>&lt;P&gt;&lt;STRONG&gt;@&lt;/STRONG&gt;&lt;A href="https://community.hortonworks.com/users/10331/mrizvi.html"&gt;Mushtaq Rizvi&lt;/A&gt;&lt;/P&gt;&lt;P&gt;As you already know, in addition to the API, Atlas uses Apache Kafka as a notification server for communication between hooks and downstream consumers of metadata notification events. There is no other Notification Server capability like SMTP. You would have to write your own filtering through events for those tables that you are interested. That is your presented option 2.&lt;/P&gt;&lt;P&gt;You may not like it, but this is the best answer as of now. If you had NiFi you could easily write that Notification Server by filtering the events based on a lookup list of tables. With latest versions of NiFi you can take advantage of powerful processors like LookupRecord, QueryRecord, also processors around SMTP, email etc.&lt;/P&gt;</description>
      <pubDate>Tue, 27 Mar 2018 22:46:04 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193376#M155440</guid>
      <dc:creator>cstanca</dc:creator>
      <dc:date>2018-03-27T22:46:04Z</dc:date>
    </item>
    <item>
      <title>Re: Tracking of Hive tables metadata changes in real time using Atlas</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193377#M155441</link>
      <description>&lt;P&gt;Thank you &lt;A rel="user" href="https://community.cloudera.com/users/3486/cstanca.html" nodeid="3486"&gt;@Constantin Stanca&lt;/A&gt; so much. This was very much needed. Can you please let me know the Nifi version compatible with HDP 2.6. I have HDP cluster already installed on google cloud, cannot install a separate HDF cluster. Is there a standalone jar of Nifi that can work in HDP cluster?&lt;/P&gt;</description>
      <pubDate>Wed, 28 Mar 2018 00:37:51 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193377#M155441</guid>
      <dc:creator>mrizvi</dc:creator>
      <dc:date>2018-03-28T00:37:51Z</dc:date>
    </item>
    <item>
      <title>Re: Tracking of Hive tables metadata changes in real time using Atlas</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193378#M155442</link>
      <description>&lt;P&gt;&lt;STRONG&gt;@&lt;/STRONG&gt;&lt;A href="https://community.hortonworks.com/users/10331/mrizvi.html"&gt;Mushtaq Rizvi&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Yes. Please follow the instructions on how to add HDF components to an existent HDP 2.6.1 cluster:&lt;/P&gt;&lt;P&gt;&lt;A href="https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.0.1/bk_installing-hdf-on-hdp/content/upgrading_ambari.html" target="_blank"&gt;https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.0.1/bk_installing-hdf-on-hdp/content/upgrading_ambari.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;This is not the latest HDF, but it is compatible with HDP 2.6.1 and I was pretty happy with its stability and recommend it.&lt;/P&gt;&lt;P&gt; You would be able to add Apache NiFi 1.5, but also Schema Registry. NiFi Registry is part of the latest HDF 3.1.x, however, you would have to install it in a separate cluster and it is not worth it the effort for what you are trying to achieve right now. I would proceed with HDP upgrade when you are ready for HDF 3.2 which will be probably launched in the next couple months. &lt;/P&gt;&lt;P&gt;In case that you can't add another node to your cluster for NiFi, try to use one of the nodes that has low CPU utilization and some disk available for NiFi lineage data storage. It depends on how much lineage you want to preserve, but you should be probably fine with several tens of GB for starters.&lt;/P&gt;&lt;P&gt;If this response helped, please vote and accept answer.&lt;/P&gt;</description>
      <pubDate>Wed, 28 Mar 2018 11:22:52 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Tracking-of-Hive-tables-metadata-changes-in-real-time-using/m-p/193378#M155442</guid>
      <dc:creator>cstanca</dc:creator>
      <dc:date>2018-03-28T11:22:52Z</dc:date>
    </item>
  </channel>
</rss>

