<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Can NiFi be used to pipe data from sql server CDC to Kafka? in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142563#M35577</link>
    <description>&lt;P&gt;The SQL Server Enterprise CDC feature can be directed to write data to a table in the SQL server, effectively keeping an audit log in a SQL table. This can be read using the NiFi QueryDatabaseTable processor. It effectively works a bit like a trigger writing and audit table which you can then read incrementally with NiFi. So, the easy answer is yes. You would just create a flow with QueryDatabaseTable (based on the timestamp as max column) -&amp;gt; PutKafka. You may need to SplitAvro between the two depending on the batch size you need to get good throughput, but that's another question &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;&lt;P&gt;That said, there are other forms of CDC which may be more efficient, but currently require third party products. Using Passive CDC, which reads the transaction logs of the SQL Server, and therefore does not put additional query load on the processor is an option here. Tools such as Attunity Replicate use this approach and write directly to Kafka. &lt;A href="https://www.brighttalk.com/webcast/9573/214447"&gt;See this webinar&lt;/A&gt; for an example of how these tools work and how you can use NiFi with them. &lt;/P&gt;</description>
    <pubDate>Wed, 27 Jul 2016 16:10:49 GMT</pubDate>
    <dc:creator>sball</dc:creator>
    <dc:date>2016-07-27T16:10:49Z</dc:date>
    <item>
      <title>Can NiFi be used to pipe data from sql server CDC to Kafka?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142561#M35575</link>
      <description />
      <pubDate>Fri, 22 Jul 2016 13:27:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142561#M35575</guid>
      <dc:creator>surajbirla</dc:creator>
      <dc:date>2016-07-22T13:27:06Z</dc:date>
    </item>
    <item>
      <title>Re: Can NiFi be used to pipe data from sql server CDC to Kafka?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142562#M35576</link>
      <description>&lt;P&gt;NiFi currently doesn't support CDC from transaction logs, but you can utilize DB triggers (hitting the target table frequently to get the change records) and publish the change db records to kafaka.&lt;/P&gt;</description>
      <pubDate>Mon, 25 Jul 2016 20:37:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142562#M35576</guid>
      <dc:creator>mpandit</dc:creator>
      <dc:date>2016-07-25T20:37:43Z</dc:date>
    </item>
    <item>
      <title>Re: Can NiFi be used to pipe data from sql server CDC to Kafka?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142563#M35577</link>
      <description>&lt;P&gt;The SQL Server Enterprise CDC feature can be directed to write data to a table in the SQL server, effectively keeping an audit log in a SQL table. This can be read using the NiFi QueryDatabaseTable processor. It effectively works a bit like a trigger writing and audit table which you can then read incrementally with NiFi. So, the easy answer is yes. You would just create a flow with QueryDatabaseTable (based on the timestamp as max column) -&amp;gt; PutKafka. You may need to SplitAvro between the two depending on the batch size you need to get good throughput, but that's another question &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;&lt;P&gt;That said, there are other forms of CDC which may be more efficient, but currently require third party products. Using Passive CDC, which reads the transaction logs of the SQL Server, and therefore does not put additional query load on the processor is an option here. Tools such as Attunity Replicate use this approach and write directly to Kafka. &lt;A href="https://www.brighttalk.com/webcast/9573/214447"&gt;See this webinar&lt;/A&gt; for an example of how these tools work and how you can use NiFi with them. &lt;/P&gt;</description>
      <pubDate>Wed, 27 Jul 2016 16:10:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Can-NiFi-be-used-to-pipe-data-from-sql-server-CDC-to-Kafka/m-p/142563#M35577</guid>
      <dc:creator>sball</dc:creator>
      <dc:date>2016-07-27T16:10:49Z</dc:date>
    </item>
  </channel>
</rss>

