<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: merge file in hdfs in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/merge-file-in-hdfs/m-p/216280#M178189</link>
    <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/45384/evaloschin.html" nodeid="45384"&gt;@eric valoschin&lt;/A&gt; the solution in the above link is not storing the output on local FS. It is streaming the output from HDFS to HDFS:&lt;/P&gt;&lt;P&gt;============================&lt;/P&gt;&lt;P&gt;A command line scriptlet to do this could be as follows:&lt;/P&gt;&lt;PRE&gt;hadoop fs -text *_fileName.txt | hadoop fs -put - targetFilename.txt
&lt;/PRE&gt;&lt;P&gt;This will cat all files that match the glob to standard output, then 
you'll pipe that stream to the put command and output the stream to an 
HDFS file named targetFilename.txt&lt;/P&gt;&lt;P&gt;=============================&lt;/P&gt;</description>
    <pubDate>Tue, 10 Oct 2017 02:58:12 GMT</pubDate>
    <dc:creator>manish1</dc:creator>
    <dc:date>2017-10-10T02:58:12Z</dc:date>
  </channel>
</rss>

