<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Nifi ListenSyslog howto ingest into HAWQ in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214047#M175963</link>
    <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;1. step we ingest SYSLOG messages with the ListenSyslog and PutHDFS in our DataLake.&lt;/P&gt;&lt;P&gt;2. step ingest with  PXF external table technology into HAWQ.&lt;/P&gt;&lt;P&gt;CREATE EXTERNAL TABLE ext_syslog_hist&lt;BR /&gt;(message TEXT)&lt;BR /&gt;LOCATION ('pxf://xxx.xxx.com:8020/apps/nifi/syslog/history/*?PROFILE=HdfsTextSimple')&lt;BR /&gt;FORMAT 'TEXT' (delimiter=E'\t');&lt;/P&gt;&lt;P&gt;Every minute there are thousands of small (100 bytes) files created in the HDFS.&lt;BR /&gt;So I'm looking for a way to ingest the incoming ListenSyslog directly into a HAWQ table.&lt;/P&gt;&lt;P&gt;Regards&lt;BR /&gt;Timo&lt;/P&gt;</description>
    <pubDate>Wed, 09 Aug 2017 01:04:00 GMT</pubDate>
    <dc:creator>timo_burmeister</dc:creator>
    <dc:date>2017-08-09T01:04:00Z</dc:date>
    <item>
      <title>Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214047#M175963</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;1. step we ingest SYSLOG messages with the ListenSyslog and PutHDFS in our DataLake.&lt;/P&gt;&lt;P&gt;2. step ingest with  PXF external table technology into HAWQ.&lt;/P&gt;&lt;P&gt;CREATE EXTERNAL TABLE ext_syslog_hist&lt;BR /&gt;(message TEXT)&lt;BR /&gt;LOCATION ('pxf://xxx.xxx.com:8020/apps/nifi/syslog/history/*?PROFILE=HdfsTextSimple')&lt;BR /&gt;FORMAT 'TEXT' (delimiter=E'\t');&lt;/P&gt;&lt;P&gt;Every minute there are thousands of small (100 bytes) files created in the HDFS.&lt;BR /&gt;So I'm looking for a way to ingest the incoming ListenSyslog directly into a HAWQ table.&lt;/P&gt;&lt;P&gt;Regards&lt;BR /&gt;Timo&lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2017 01:04:00 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214047#M175963</guid>
      <dc:creator>timo_burmeister</dc:creator>
      <dc:date>2017-08-09T01:04:00Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214048#M175964</link>
      <description>&lt;P&gt;You could use PutSQL to issue INSERT statements directly against HAWQ. Assign the data item values to attributes, as described in the &lt;A href="https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-standard-nar/1.3.0/org.apache.nifi.processors.standard.PutSQL/index.html"&gt;PutSQL documentation&lt;/A&gt;. Then put the INSERT statement in the content of the Flow File, with '?' placeholders for the individual values.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2017 01:13:45 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214048#M175964</guid>
      <dc:creator>Former Member</dc:creator>
      <dc:date>2017-08-09T01:13:45Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214049#M175965</link>
      <description>&lt;P&gt;You could also use MergeContent before PutHDFS so that you can merge together all of the small files based on size or time before writing to HDFS.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2017 02:47:30 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214049#M175965</guid>
      <dc:creator>bbende</dc:creator>
      <dc:date>2017-08-09T02:47:30Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214050#M175966</link>
      <description>&lt;P&gt;You definitely do not want to execute an INSERT for every row of data.  That will be really, really slow.  &lt;/P&gt;&lt;P&gt;I suggest you use the gpfdist protocol to load the log files directly into HAWQ.  You start a gpfdist process where the files are located, create an external table that uses the gpfdist location, and then execute a single INSERT statement to load all of the data from the files into your HAWQ table.  &lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2017 03:18:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214050#M175966</guid>
      <dc:creator>jroberts</dc:creator>
      <dc:date>2017-08-09T03:18:19Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214051#M175967</link>
      <description>&lt;P&gt;Thanks,&lt;BR /&gt;this was really easy. &lt;BR /&gt;I just generating "insert into table" statements including the Flow File with the ReplaceText processor and connect it with the PutSQL processor.&lt;BR /&gt;&lt;BR /&gt;regards&lt;BR /&gt;Timo&lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2017 23:02:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214051#M175967</guid>
      <dc:creator>timo_burmeister</dc:creator>
      <dc:date>2017-08-09T23:02:56Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214052#M175968</link>
      <description>&lt;P&gt;Is this equivalent to: INSERT INTO &amp;lt;table&amp;gt; (col1, col2, col3) VALUES (?, ?, ?);&lt;/P&gt;</description>
      <pubDate>Thu, 10 Aug 2017 02:18:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214052#M175968</guid>
      <dc:creator>jroberts</dc:creator>
      <dc:date>2017-08-10T02:18:25Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214053#M175969</link>
      <description>&lt;P&gt;The PutSQL option is by far the worst option.  This is the wrong answer.  Please, don't do that.  Hellmer clearly doesn't use HAWQ.&lt;/P&gt;&lt;P&gt;I wrote a blog post over a year ago on this subject.  The singleton INSERT solution that is common for OLTP databases is the absolute worst solution for HAWQ.  In other words, "INSERT INTO table_name (columns...) VALUES (...);"  &lt;/P&gt;&lt;P&gt;Here is my blog post covering loading data into HAWQ: &lt;A href="http://www.pivotalguru.com/?p=1425" target="_blank"&gt;http://www.pivotalguru.com/?p=1425&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 10 Aug 2017 02:51:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214053#M175969</guid>
      <dc:creator>jroberts</dc:creator>
      <dc:date>2017-08-10T02:51:23Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214054#M175970</link>
      <description>&lt;P&gt;Hi Bryan,&lt;/P&gt;&lt;P&gt;how can i merge the FlowData Files  to a multiline File?&lt;BR /&gt;Using the MergeCentent processor i got a concatenated line.&lt;BR /&gt;input: &lt;/P&gt;&lt;P&gt;1 FlowData: &amp;lt;30&amp;gt;Aug 11 06:27:26 xxx.xxx.com systemd[28892]: Stopping Timers.&lt;/P&gt;&lt;P&gt;2 FlowData:&amp;lt;30&amp;gt;Aug 11 06:27:15 xxx.xxx.com systemd[24517]: Stopping Paths&lt;/P&gt;&lt;P&gt;etc...&lt;/P&gt;&lt;P&gt;MergeContent output:&lt;/P&gt;&lt;P&gt;&amp;lt;30&amp;gt;Aug 11 06:27:26 xxx.xxx.com systemd[28892]: Stopping Timers.&amp;lt;30&amp;gt;Aug 11 06:27:15 xxx.xxx.com systemd[24517]:Stopping Paths&lt;/P&gt;&lt;P&gt;I need the following multiline structure:&lt;/P&gt;&lt;P&gt;&amp;lt;30&amp;gt;Aug 11 06:27:26 xxx.xxx.com systemd[28892]: Stopping Timers.&lt;BR /&gt;&amp;lt;30&amp;gt;Aug 11 06:27:15 xxx.xxx.com systemd[24517]:Stopping Paths&lt;BR /&gt;&lt;BR /&gt;any solution?&lt;BR /&gt;&lt;BR /&gt;regards&lt;BR /&gt;Timo&lt;/P&gt;</description>
      <pubDate>Fri, 11 Aug 2017 18:52:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214054#M175970</guid>
      <dc:creator>timo_burmeister</dc:creator>
      <dc:date>2017-08-11T18:52:59Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214055#M175971</link>
      <description>&lt;P&gt;Just set the "Delimiter Strategy" in MergeContent to "Text" and then set the "Demarcator" property to shift+enter to make a new line. It will insert the demarcator between each flow file that it merges.&lt;/P&gt;</description>
      <pubDate>Fri, 11 Aug 2017 19:22:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214055#M175971</guid>
      <dc:creator>bbende</dc:creator>
      <dc:date>2017-08-11T19:22:21Z</dc:date>
    </item>
    <item>
      <title>Re: Nifi ListenSyslog howto ingest into HAWQ</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214056#M175972</link>
      <description>&lt;P&gt;Great, works !!!&lt;BR /&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Fri, 11 Aug 2017 20:44:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Nifi-ListenSyslog-howto-ingest-into-HAWQ/m-p/214056#M175972</guid>
      <dc:creator>timo_burmeister</dc:creator>
      <dc:date>2017-08-11T20:44:27Z</dc:date>
    </item>
  </channel>
</rss>

