<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Sqoop Import to Hive with Compression in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116836#M79630</link>
    <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1496/nileshpatil.html" nodeid="1496"&gt;@Nilesh&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Below given is your solution:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Input:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;mysql&amp;gt; select * from SERDES;&lt;/BLOCKQUOTE&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;| SERDE_ID | NAME | SLIB 
|&lt;/P&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;|  56 | NULL |
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |&lt;/P&gt;&lt;P&gt;|  57 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  58 | NULL |
NULL 
|&lt;/P&gt;&lt;P&gt;|  59 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  60 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  61 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  62 | NULL |
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |&lt;/P&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;7 rows in set (0.00 sec)&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;&lt;P&gt;sqoop import --connect jdbc:mysql://test:3306/hive \&lt;/P&gt;&lt;P&gt; 
--username hive \&lt;/P&gt;&lt;P&gt; 
--password test \&lt;/P&gt;&lt;P&gt;  --table
SERDES \&lt;/P&gt;&lt;P&gt; 
--hcatalog-database test \&lt;/P&gt;&lt;P&gt; 
--hcatalog-table SERDES \&lt;/P&gt;&lt;P&gt; 
--create-hcatalog-table \&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt; 
--hcatalog-storage-stanza "stored as orcfile" \&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;  --outdir
sqoop_import \&lt;/P&gt;&lt;P&gt;  -m 1 \&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt; 
--compression-codec org.apache.hadoop.io.compress.SnappyCodec \&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;    --driver com.mysql.jdbc.Driver&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;STRONG&gt;Logs:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;16/05/06 13:30:46 INFO hcat.SqoopHCatUtilities: HCatalog
Create table statement:&lt;/P&gt;&lt;P&gt;create table `demand_db`.`serdes` (&lt;/P&gt;&lt;P&gt;  `serde_id`
bigint,&lt;/P&gt;&lt;P&gt;  `name`
varchar(128),&lt;/P&gt;&lt;P&gt;  `slib`
varchar(4000))&lt;/P&gt;&lt;P&gt;stored as orcfile&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;16/05/06 13:32:55 INFO mapreduce.Job: Job
job_1462201699379_0089 running in uber mode : false&lt;/P&gt;&lt;P&gt;16/05/06 13:32:55 INFO mapreduce.Job:  map 0% reduce 0%&lt;/P&gt;&lt;P&gt;16/05/06 13:33:07 INFO mapreduce.Job:  map 100% reduce 0%&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.Job: Job
job_1462201699379_0089 completed successfully&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.Job: Counters: 30&lt;/P&gt;&lt;P&gt;  File System
Counters&lt;/P&gt;&lt;P&gt;  FILE:
Number of bytes read=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of bytes written=297179&lt;/P&gt;&lt;P&gt;  FILE:
Number of read operations=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of large read operations=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of write operations=0&lt;/P&gt;&lt;P&gt;  HDFS:
Number of bytes read=87&lt;/P&gt;&lt;P&gt;  HDFS:
Number of bytes written=676&lt;/P&gt;&lt;P&gt;  HDFS:
Number of read operations=4&lt;/P&gt;&lt;P&gt;  HDFS:
Number of large read operations=0&lt;/P&gt;&lt;P&gt;  HDFS:
Number of write operations=2&lt;/P&gt;&lt;P&gt;  Job
Counters&lt;/P&gt;&lt;P&gt;  Launched
map tasks=1&lt;/P&gt;&lt;P&gt;  Other
local map tasks=1&lt;/P&gt;&lt;P&gt;  Total
time spent by all maps in occupied slots (ms)=14484&lt;/P&gt;&lt;P&gt;  Total
time spent by all reduces in occupied slots (ms)=0&lt;/P&gt;&lt;P&gt;  Total
time spent by all map tasks (ms)=7242&lt;/P&gt;&lt;P&gt;  Total
vcore-seconds taken by all map tasks=7242&lt;/P&gt;&lt;P&gt;  Total
megabyte-seconds taken by all map tasks=11123712&lt;/P&gt;&lt;P&gt;  Map-Reduce
Framework&lt;/P&gt;&lt;P&gt;  Map
input records=8&lt;/P&gt;&lt;P&gt;  Map
output records=8&lt;/P&gt;&lt;P&gt;  Input
split bytes=87&lt;/P&gt;&lt;P&gt;  Spilled
Records=0&lt;/P&gt;&lt;P&gt;  Failed
Shuffles=0&lt;/P&gt;&lt;P&gt;  Merged
Map outputs=0&lt;/P&gt;&lt;P&gt;  GC
time elapsed (ms)=92&lt;/P&gt;&lt;P&gt;  CPU
time spent (ms)=4620&lt;/P&gt;&lt;P&gt;  Physical
memory (bytes) snapshot=353759232&lt;/P&gt;&lt;P&gt;  Virtual
memory (bytes) snapshot=3276144640&lt;/P&gt;&lt;P&gt;  Total
committed heap usage (bytes)=175112192&lt;/P&gt;&lt;P&gt;  File Input
Format Counters&lt;/P&gt;&lt;P&gt;  Bytes
Read=0&lt;/P&gt;&lt;P&gt;  File Output
Format Counters&lt;/P&gt;&lt;P&gt;  Bytes
Written=0&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.ImportJobBase: Transferred
676 bytes in 130.8366 seconds (5.1668 bytes/sec)&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.ImportJobBase: Retrieved 8
records.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Output:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;hive&amp;gt; select * from serdes;&lt;/BLOCKQUOTE&gt;&lt;P&gt;OK&lt;/P&gt;&lt;P&gt;56  NULL  org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe&lt;/P&gt;&lt;P&gt;57  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;58  NULL  NULL&lt;/P&gt;&lt;P&gt;59  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;60  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;61  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;62  NULL  org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe&lt;/P&gt;&lt;P&gt;63  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;Time taken: 2.711 seconds, Fetched: 8 row(s)&lt;/P&gt;&lt;P&gt;hive&amp;gt;&lt;/P&gt;</description>
    <pubDate>Sat, 07 May 2016 00:41:10 GMT</pubDate>
    <dc:creator>rbiswas1</dc:creator>
    <dc:date>2016-05-07T00:41:10Z</dc:date>
    <item>
      <title>Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116830#M79624</link>
      <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;How to import oracle tables in hive with compressed format via sqoop command. &lt;/P&gt;&lt;P&gt;Where to configure compression. i.e  MapReduce or HDFS or Hive.&lt;/P&gt;&lt;P&gt;Should I configure compression for Mapreduce ,HDFS and Hive.&lt;/P&gt;&lt;P&gt;Which is the best compression formats supported by Hive.&lt;/P&gt;</description>
      <pubDate>Tue, 26 Apr 2016 21:37:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116830#M79624</guid>
      <dc:creator>Nilesh</dc:creator>
      <dc:date>2016-04-26T21:37:01Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116831#M79625</link>
      <description>&lt;P&gt; &lt;A rel="user" href="https://community.cloudera.com/users/1496/nileshpatil.html" nodeid="1496"&gt;@Nilesh&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Below is an example where I imported a table that in mysql to hive in ORC format. You don't need to create the ORC backed table in advance. key is --hcatalog-storage-stanza. &lt;/P&gt;&lt;PRE&gt;sqoop import --connect "jdbc:mysql://sandbox.hortonworks.com/hive" --driver com.mysql.jdbc.Driver --username hive --password hive --table testtable --hcatalog-database default --hcatalog-table testtable  --create-hcatalog-table --hcatalog-storage-stanza "stored as orcfile"  -m 1&lt;/PRE&gt;</description>
      <pubDate>Tue, 26 Apr 2016 21:44:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116831#M79625</guid>
      <dc:creator>ravi1</dc:creator>
      <dc:date>2016-04-26T21:44:16Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116832#M79626</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/216/ravi.html" nodeid="216"&gt;@Ravi Mutyala&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Does sqoop import support orc file. Where is compressed parameter.&lt;/P&gt;</description>
      <pubDate>Wed, 27 Apr 2016 21:11:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116832#M79626</guid>
      <dc:creator>Nilesh</dc:creator>
      <dc:date>2016-04-27T21:11:41Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116833#M79627</link>
      <description>&lt;P&gt;Above example shows you that. --hcatalog-storage_stanza "stored as orcfile" will create a new hive table through hcatalog and data is stored as ORC. You can also create a hive table with ORC format and then use --hcatalog-table to write directly as ORC. &lt;/P&gt;</description>
      <pubDate>Wed, 27 Apr 2016 22:17:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116833#M79627</guid>
      <dc:creator>ravi1</dc:creator>
      <dc:date>2016-04-27T22:17:35Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116834#M79628</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/216/ravi.html" nodeid="216"&gt;@Ravi Mutyala&lt;/A&gt;&lt;P&gt;Does it compressed the output.&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 13:10:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116834#M79628</guid>
      <dc:creator>Nilesh</dc:creator>
      <dc:date>2016-04-29T13:10:06Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116835#M79629</link>
      <description>&lt;P&gt;Yes. Output will be ORC file(s).&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 20:17:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116835#M79629</guid>
      <dc:creator>ravi1</dc:creator>
      <dc:date>2016-04-29T20:17:55Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116836#M79630</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1496/nileshpatil.html" nodeid="1496"&gt;@Nilesh&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Below given is your solution:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Input:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;mysql&amp;gt; select * from SERDES;&lt;/BLOCKQUOTE&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;| SERDE_ID | NAME | SLIB 
|&lt;/P&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;|  56 | NULL |
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |&lt;/P&gt;&lt;P&gt;|  57 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  58 | NULL |
NULL 
|&lt;/P&gt;&lt;P&gt;|  59 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  60 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  61 | NULL |
org.apache.hadoop.hive.ql.io.orc.OrcSerde  |&lt;/P&gt;&lt;P&gt;|  62 | NULL |
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |&lt;/P&gt;&lt;P&gt;+----------+------+----------------------------------------------------+&lt;/P&gt;&lt;P&gt;7 rows in set (0.00 sec)&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;&lt;P&gt;sqoop import --connect jdbc:mysql://test:3306/hive \&lt;/P&gt;&lt;P&gt; 
--username hive \&lt;/P&gt;&lt;P&gt; 
--password test \&lt;/P&gt;&lt;P&gt;  --table
SERDES \&lt;/P&gt;&lt;P&gt; 
--hcatalog-database test \&lt;/P&gt;&lt;P&gt; 
--hcatalog-table SERDES \&lt;/P&gt;&lt;P&gt; 
--create-hcatalog-table \&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt; 
--hcatalog-storage-stanza "stored as orcfile" \&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;  --outdir
sqoop_import \&lt;/P&gt;&lt;P&gt;  -m 1 \&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt; 
--compression-codec org.apache.hadoop.io.compress.SnappyCodec \&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;    --driver com.mysql.jdbc.Driver&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;STRONG&gt;Logs:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;16/05/06 13:30:46 INFO hcat.SqoopHCatUtilities: HCatalog
Create table statement:&lt;/P&gt;&lt;P&gt;create table `demand_db`.`serdes` (&lt;/P&gt;&lt;P&gt;  `serde_id`
bigint,&lt;/P&gt;&lt;P&gt;  `name`
varchar(128),&lt;/P&gt;&lt;P&gt;  `slib`
varchar(4000))&lt;/P&gt;&lt;P&gt;stored as orcfile&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;...&lt;/P&gt;&lt;P&gt;16/05/06 13:32:55 INFO mapreduce.Job: Job
job_1462201699379_0089 running in uber mode : false&lt;/P&gt;&lt;P&gt;16/05/06 13:32:55 INFO mapreduce.Job:  map 0% reduce 0%&lt;/P&gt;&lt;P&gt;16/05/06 13:33:07 INFO mapreduce.Job:  map 100% reduce 0%&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.Job: Job
job_1462201699379_0089 completed successfully&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.Job: Counters: 30&lt;/P&gt;&lt;P&gt;  File System
Counters&lt;/P&gt;&lt;P&gt;  FILE:
Number of bytes read=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of bytes written=297179&lt;/P&gt;&lt;P&gt;  FILE:
Number of read operations=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of large read operations=0&lt;/P&gt;&lt;P&gt;  FILE:
Number of write operations=0&lt;/P&gt;&lt;P&gt;  HDFS:
Number of bytes read=87&lt;/P&gt;&lt;P&gt;  HDFS:
Number of bytes written=676&lt;/P&gt;&lt;P&gt;  HDFS:
Number of read operations=4&lt;/P&gt;&lt;P&gt;  HDFS:
Number of large read operations=0&lt;/P&gt;&lt;P&gt;  HDFS:
Number of write operations=2&lt;/P&gt;&lt;P&gt;  Job
Counters&lt;/P&gt;&lt;P&gt;  Launched
map tasks=1&lt;/P&gt;&lt;P&gt;  Other
local map tasks=1&lt;/P&gt;&lt;P&gt;  Total
time spent by all maps in occupied slots (ms)=14484&lt;/P&gt;&lt;P&gt;  Total
time spent by all reduces in occupied slots (ms)=0&lt;/P&gt;&lt;P&gt;  Total
time spent by all map tasks (ms)=7242&lt;/P&gt;&lt;P&gt;  Total
vcore-seconds taken by all map tasks=7242&lt;/P&gt;&lt;P&gt;  Total
megabyte-seconds taken by all map tasks=11123712&lt;/P&gt;&lt;P&gt;  Map-Reduce
Framework&lt;/P&gt;&lt;P&gt;  Map
input records=8&lt;/P&gt;&lt;P&gt;  Map
output records=8&lt;/P&gt;&lt;P&gt;  Input
split bytes=87&lt;/P&gt;&lt;P&gt;  Spilled
Records=0&lt;/P&gt;&lt;P&gt;  Failed
Shuffles=0&lt;/P&gt;&lt;P&gt;  Merged
Map outputs=0&lt;/P&gt;&lt;P&gt;  GC
time elapsed (ms)=92&lt;/P&gt;&lt;P&gt;  CPU
time spent (ms)=4620&lt;/P&gt;&lt;P&gt;  Physical
memory (bytes) snapshot=353759232&lt;/P&gt;&lt;P&gt;  Virtual
memory (bytes) snapshot=3276144640&lt;/P&gt;&lt;P&gt;  Total
committed heap usage (bytes)=175112192&lt;/P&gt;&lt;P&gt;  File Input
Format Counters&lt;/P&gt;&lt;P&gt;  Bytes
Read=0&lt;/P&gt;&lt;P&gt;  File Output
Format Counters&lt;/P&gt;&lt;P&gt;  Bytes
Written=0&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.ImportJobBase: Transferred
676 bytes in 130.8366 seconds (5.1668 bytes/sec)&lt;/P&gt;&lt;P&gt;16/05/06 13:33:09 INFO mapreduce.ImportJobBase: Retrieved 8
records.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Output:&lt;/STRONG&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;hive&amp;gt; select * from serdes;&lt;/BLOCKQUOTE&gt;&lt;P&gt;OK&lt;/P&gt;&lt;P&gt;56  NULL  org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe&lt;/P&gt;&lt;P&gt;57  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;58  NULL  NULL&lt;/P&gt;&lt;P&gt;59  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;60  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;61  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;62  NULL  org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe&lt;/P&gt;&lt;P&gt;63  NULL  org.apache.hadoop.hive.ql.io.orc.OrcSerde&lt;/P&gt;&lt;P&gt;Time taken: 2.711 seconds, Fetched: 8 row(s)&lt;/P&gt;&lt;P&gt;hive&amp;gt;&lt;/P&gt;</description>
      <pubDate>Sat, 07 May 2016 00:41:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116836#M79630</guid>
      <dc:creator>rbiswas1</dc:creator>
      <dc:date>2016-05-07T00:41:10Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116837#M79631</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1496/nileshpatil.html" nodeid="1496"&gt;@Nilesh&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Did it worked?&lt;/P&gt;</description>
      <pubDate>Wed, 11 May 2016 11:00:42 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116837#M79631</guid>
      <dc:creator>rbiswas1</dc:creator>
      <dc:date>2016-05-11T11:00:42Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116838#M79632</link>
      <description>&lt;P&gt;A late answer but maybe it will help someone &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;&lt;P&gt;I am just adding to what @&lt;A href="https://community.hortonworks.com/users/216/ravi.html"&gt;Ravi Mutyala&lt;/A&gt; has mentioned :&lt;/P&gt;&lt;PRE&gt;sqoop import --null-string '\\N' --null-non-string '\\N' --hive-delims-replacement '\0D' --num-mappers 8 --validate --hcatalog-home /usr/hdp/current/hive-webhcat --hcatalog-database default --hcatalog-table Inactivity --create-hcatalog-table --hcatalog-storage-stanza 'stored as orc tblproperties ("orc.compress"="ZLIB")' --connect 'jdbc:sqlserver://&amp;lt;IP&amp;gt;;database=&amp;lt;db-name&amp;gt;' --username --password --table Inactivity -- --schema QlikView 2&amp;gt;&amp;amp;1| tee -a log&lt;/PRE&gt;&lt;P&gt;Now if you describe the table :&lt;/P&gt;&lt;PRE&gt;0: jdbc:hive2://&amp;gt; describe formatted inactivity;
OK
16/08/18 11:23:25 [main]: WARN lazy.LazyStruct: Extra bytes detected at the end of the row! Ignoring similar problems.
+-------------------------------+----------------------------------------------------------------------+-----------------------+--+
|           col_name            |                              data_type                               |        comment        |
+-------------------------------+----------------------------------------------------------------------+-----------------------+--+
| # col_name                    | data_type                                                            | comment               |
|                               | NULL                                                                 | NULL                  |
| period                        | int                                                                  |                       |
| vin                           | string                                                               |                       |
| customerid                    | int                                                                  |                       |
| subscriberdealersisid         | string                                                               |                       |
| subscriberdistributorsisid    | string                                                               |                       |
| packagename                   | string                                                               |                       |
| timemodify                    | string                                                               |                       |
|                               | NULL                                                                 | NULL                  |
| # Detailed Table Information  | NULL                                                                 | NULL                  |
| Database:                     | default                                                              | NULL                  |
| Owner:                        | hive                                                                 | NULL                  |
| CreateTime:                   | Thu Aug 18 11:20:28 CEST 2016                                        | NULL                  |
| LastAccessTime:               | UNKNOWN                                                              | NULL                  |
| Protect Mode:                 | None                                                                 | NULL                  |
| Retention:                    | 0                                                                    | NULL                  |
| Location:                     | hdfs://l4283t.sss.com:8020/apps/hive/warehouse/inactivity  | NULL                  |
| Table Type:                   | MANAGED_TABLE                                                        | NULL                  |
| Table Parameters:             | NULL                                                                 | NULL                  |
|                               | orc.compress                                                         | ZLIB                  |
|                               | transient_lastDdlTime                                                | 1471512028            |
|                               | NULL                                                                 | NULL                  |
| # Storage Information         | NULL                                                                 | NULL                  |
| SerDe Library:                | org.apache.hadoop.hive.ql.io.orc.OrcSerde                            | NULL                  |
| InputFormat:                  | org.apache.hadoop.hive.ql.io.orc.OrcInputFormat                      | NULL                  |
| OutputFormat:                 | org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat                     | NULL                  |
| Compressed:                   | No                                                                   | NULL                  |
| Num Buckets:                  | -1                                                                   | NULL                  |
| Bucket Columns:               | []                                                                   | NULL                  |
| Sort Columns:                 | []                                                                   | NULL                  |
| Storage Desc Params:          | NULL                                                                 | NULL                  |
|                               | serialization.format                                                 | 1                     |
+-------------------------------+----------------------------------------------------------------------+-----------------------+--+
33 rows selected (0.425 seconds)&lt;/PRE&gt;&lt;P&gt;To verify if the compression has really taken place, you can first import the table without any compression and execute &lt;/P&gt;&lt;PRE&gt;analyze table &amp;lt;table-name&amp;gt;compute statistics&lt;/PRE&gt;&lt;P&gt;and note the 'totalSize'. Then repeat the process with compression and compare the 'totalSize'&lt;/P&gt;</description>
      <pubDate>Thu, 18 Aug 2016 18:06:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116838#M79632</guid>
      <dc:creator>kaliyugantagoni</dc:creator>
      <dc:date>2016-08-18T18:06:20Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop Import to Hive with Compression</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116839#M79633</link>
      <description>&lt;P&gt;For a comparison between compression formats take a look at this link:&lt;/P&gt;&lt;P&gt;&lt;A href="http://comphadoop.weebly.com/" target="_blank"&gt;http://comphadoop.weebly.com/&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 24 Jul 2017 21:42:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Sqoop-Import-to-Hive-with-Compression/m-p/116839#M79633</guid>
      <dc:creator>egarelnabi</dc:creator>
      <dc:date>2017-07-24T21:42:38Z</dc:date>
    </item>
  </channel>
</rss>

