<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Getting Error in Sqoop Import from Oracle &amp; Mssql database in Hive Table as Parquet Format in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Getting-Error-in-Sqoop-Import-from-Oracle-Mssql-database-in/m-p/132691#M23192</link>
    <description>&lt;P&gt;I am getting same error in using sqoop import while fetching data from either Mssql or Oracle database. &lt;/P&gt;&lt;P&gt;Oracle -&amp;gt;
&lt;/P&gt;&lt;P&gt;sqoop import --connect "jdbc:oracle:thin:@&amp;lt;ServerName&amp;gt;:&amp;lt;PortName&amp;gt;:&amp;lt;Database&amp;gt;" --hadoop-home "/usr/hdp/2.4.0.0-169/hadoop" --username &amp;lt;UserName&amp;gt; --password &amp;lt;Password&amp;gt; --table &amp;lt;TableName&amp;gt; --hive-overwrite --columns "COL_NAM" --hive-import -hive-database ia_db -hive-table par_str_cols --map-column-hive db_user_nam=string --as-parquetfile -m 1&lt;/P&gt;&lt;P&gt;Mssql -&amp;gt;&lt;/P&gt;&lt;P&gt;sqoop import --connect "jdbc:sqlserver://&amp;lt;ServerName&amp;gt;:&amp;lt;PortNo&amp;gt;;database=&amp;lt;DatabaseName&amp;gt;" --username &amp;lt;UserName&amp;gt; --password &amp;lt;Password&amp;gt; --table &amp;lt;TableName&amp;gt; --hive-overwrite --columns "COLUMN1, COLUMN2" --where "COLUMN1 = 7390" --hive-import --target-dir /apps/hive/warehouse/ia_db.db/par_int_cols --as-parquetfile -m 1&lt;/P&gt;&lt;P&gt;I have tried with Integer, Decimal and String columns and getting same error. &lt;/P&gt;&lt;P&gt;Note - I am able to import same data in Hive tables as simple text files but getting this error while importing in Parquet format. &lt;/P&gt;&lt;P&gt;Can someone give some pointers on what "Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V" is this?&lt;/P&gt;&lt;P&gt;**********************************************************************************************************&lt;/P&gt;&lt;P&gt;16/03/17 10:43:03 INFO hive.metastore: Trying to connect to metastore with URI thrift://sandbox.hortonworks.com:9083
16/03/17 10:43:03 INFO hive.metastore: Connected to metastore. &lt;/P&gt;&lt;P&gt;16/03/17 10:43:05 INFO impl.TimelineClientImpl: Timeline service address: &lt;A href="http://sandbox.hortonworks.com:8188/ws/v1/timeline/" target="_blank"&gt;http://sandbox.hortonworks.com:8188/ws/v1/timeline/&lt;/A&gt; &lt;/P&gt;&lt;P&gt;16/03/17 10:43:05 INFO client.RMProxy: Connecting to ResourceManager at sandbox.hortonworks.com/10.0.2.15:8050
16/03/17 10:43:05 INFO hive.metastore: Trying to connect to metastore with URI thrift://sandbox.hortonworks.com:9083
16/03/17 10:43:05 INFO hive.metastore: Connected to metastore. &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO db.DBInputFormat: Using read commited transaction isolation &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO mapreduce.JobSubmitter: number of splits:1 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO impl.YarnClientImpl: Submitted application application_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO mapreduce.Job: The url to track the job: &lt;A href="http://sandbox.hortonworks.com:8088/proxy/application_1458111880471_0054/" target="_blank"&gt;http://sandbox.hortonworks.com:8088/proxy/application_1458111880471_0054/&lt;/A&gt; &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO mapreduce.Job: Running job: job_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:28 INFO mapreduce.Job: Job job_1458111880471_0054 running in uber mode : false &lt;/P&gt;&lt;P&gt;16/03/17 10:43:28 INFO mapreduce.Job:  map 0% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:42 INFO mapreduce.Job:  map 100% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:42 INFO mapreduce.Job: Task Id : attempt_1458111880471_0054_m_000000_0, Status : FAILED &lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V &lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;Container killed by the ApplicationMaster. &lt;/P&gt;&lt;P&gt;Container killed on request. Exit code is 143 &lt;/P&gt;&lt;P&gt;Container exited with a non-zero exit code 143 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:43 INFO mapreduce.Job:  map 0% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:56 INFO mapreduce.Job:  map 100% reduce 0%
16/03/17 10:43:56 &lt;/P&gt;&lt;P&gt;INFO mapreduce.Job: Task Id : attempt_1458111880471_0054_m_000000_1, Status : FAILED &lt;/P&gt;&lt;P&gt;Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V &lt;/P&gt;&lt;P&gt;Container killed by the ApplicationMaster. &lt;/P&gt;&lt;P&gt;Container killed on request. Exit code is 143 &lt;/P&gt;&lt;P&gt;Container exited with a non-zero exit code 143&lt;/P&gt;&lt;P&gt;**********************************************************************************************************************&lt;/P&gt;</description>
    <pubDate>Thu, 17 Mar 2016 19:01:14 GMT</pubDate>
    <dc:creator>anjali_sud</dc:creator>
    <dc:date>2016-03-17T19:01:14Z</dc:date>
    <item>
      <title>Getting Error in Sqoop Import from Oracle &amp; Mssql database in Hive Table as Parquet Format</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Getting-Error-in-Sqoop-Import-from-Oracle-Mssql-database-in/m-p/132691#M23192</link>
      <description>&lt;P&gt;I am getting same error in using sqoop import while fetching data from either Mssql or Oracle database. &lt;/P&gt;&lt;P&gt;Oracle -&amp;gt;
&lt;/P&gt;&lt;P&gt;sqoop import --connect "jdbc:oracle:thin:@&amp;lt;ServerName&amp;gt;:&amp;lt;PortName&amp;gt;:&amp;lt;Database&amp;gt;" --hadoop-home "/usr/hdp/2.4.0.0-169/hadoop" --username &amp;lt;UserName&amp;gt; --password &amp;lt;Password&amp;gt; --table &amp;lt;TableName&amp;gt; --hive-overwrite --columns "COL_NAM" --hive-import -hive-database ia_db -hive-table par_str_cols --map-column-hive db_user_nam=string --as-parquetfile -m 1&lt;/P&gt;&lt;P&gt;Mssql -&amp;gt;&lt;/P&gt;&lt;P&gt;sqoop import --connect "jdbc:sqlserver://&amp;lt;ServerName&amp;gt;:&amp;lt;PortNo&amp;gt;;database=&amp;lt;DatabaseName&amp;gt;" --username &amp;lt;UserName&amp;gt; --password &amp;lt;Password&amp;gt; --table &amp;lt;TableName&amp;gt; --hive-overwrite --columns "COLUMN1, COLUMN2" --where "COLUMN1 = 7390" --hive-import --target-dir /apps/hive/warehouse/ia_db.db/par_int_cols --as-parquetfile -m 1&lt;/P&gt;&lt;P&gt;I have tried with Integer, Decimal and String columns and getting same error. &lt;/P&gt;&lt;P&gt;Note - I am able to import same data in Hive tables as simple text files but getting this error while importing in Parquet format. &lt;/P&gt;&lt;P&gt;Can someone give some pointers on what "Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V" is this?&lt;/P&gt;&lt;P&gt;**********************************************************************************************************&lt;/P&gt;&lt;P&gt;16/03/17 10:43:03 INFO hive.metastore: Trying to connect to metastore with URI thrift://sandbox.hortonworks.com:9083
16/03/17 10:43:03 INFO hive.metastore: Connected to metastore. &lt;/P&gt;&lt;P&gt;16/03/17 10:43:05 INFO impl.TimelineClientImpl: Timeline service address: &lt;A href="http://sandbox.hortonworks.com:8188/ws/v1/timeline/" target="_blank"&gt;http://sandbox.hortonworks.com:8188/ws/v1/timeline/&lt;/A&gt; &lt;/P&gt;&lt;P&gt;16/03/17 10:43:05 INFO client.RMProxy: Connecting to ResourceManager at sandbox.hortonworks.com/10.0.2.15:8050
16/03/17 10:43:05 INFO hive.metastore: Trying to connect to metastore with URI thrift://sandbox.hortonworks.com:9083
16/03/17 10:43:05 INFO hive.metastore: Connected to metastore. &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO db.DBInputFormat: Using read commited transaction isolation &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO mapreduce.JobSubmitter: number of splits:1 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:11 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO impl.YarnClientImpl: Submitted application application_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO mapreduce.Job: The url to track the job: &lt;A href="http://sandbox.hortonworks.com:8088/proxy/application_1458111880471_0054/" target="_blank"&gt;http://sandbox.hortonworks.com:8088/proxy/application_1458111880471_0054/&lt;/A&gt; &lt;/P&gt;&lt;P&gt;16/03/17 10:43:13 INFO mapreduce.Job: Running job: job_1458111880471_0054 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:28 INFO mapreduce.Job: Job job_1458111880471_0054 running in uber mode : false &lt;/P&gt;&lt;P&gt;16/03/17 10:43:28 INFO mapreduce.Job:  map 0% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:42 INFO mapreduce.Job:  map 100% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:42 INFO mapreduce.Job: Task Id : attempt_1458111880471_0054_m_000000_0, Status : FAILED &lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V &lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;Container killed by the ApplicationMaster. &lt;/P&gt;&lt;P&gt;Container killed on request. Exit code is 143 &lt;/P&gt;&lt;P&gt;Container exited with a non-zero exit code 143 &lt;/P&gt;&lt;P&gt;16/03/17 10:43:43 INFO mapreduce.Job:  map 0% reduce 0% &lt;/P&gt;&lt;P&gt;16/03/17 10:43:56 INFO mapreduce.Job:  map 100% reduce 0%
16/03/17 10:43:56 &lt;/P&gt;&lt;P&gt;INFO mapreduce.Job: Task Id : attempt_1458111880471_0054_m_000000_1, Status : FAILED &lt;/P&gt;&lt;P&gt;Error: parquet.Preconditions.checkArgument(ZLjava/lang/String;[Ljava/lang/Object;)V &lt;/P&gt;&lt;P&gt;Container killed by the ApplicationMaster. &lt;/P&gt;&lt;P&gt;Container killed on request. Exit code is 143 &lt;/P&gt;&lt;P&gt;Container exited with a non-zero exit code 143&lt;/P&gt;&lt;P&gt;**********************************************************************************************************************&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2016 19:01:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Getting-Error-in-Sqoop-Import-from-Oracle-Mssql-database-in/m-p/132691#M23192</guid>
      <dc:creator>anjali_sud</dc:creator>
      <dc:date>2016-03-17T19:01:14Z</dc:date>
    </item>
    <item>
      <title>Re: Getting Error in Sqoop Import from Oracle &amp; Mssql database in Hive Table as Parquet Format</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Getting-Error-in-Sqoop-Import-from-Oracle-Mssql-database-in/m-p/132692#M23193</link>
      <description>&lt;P&gt;I was able to reproduce this issue and it looks like a jar version mismatch for parquet on sandbox. Is there a reason for using parquet instead of ORC here? While both are supported on hive, ORC has advantages when using with hive since some of the stinger initiative improvements to hive take advantage of ORC. &lt;/P&gt;&lt;P&gt;Here is an example of using ORC from sqoop. &lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;sqoop import --connect "jdbc:sqlserver://&amp;lt;ServerName&amp;gt;:&amp;lt;PortNo&amp;gt;;database=&amp;lt;DatabaseName&amp;gt;" --username &amp;lt;UserName&amp;gt; --password &amp;lt;Password&amp;gt; --table &amp;lt;TableName&amp;gt;  --columns "COLUMN1, COLUMN2" --where "COLUMN1 = 7390" --hcatalog-database default --hcatalog-table my_table_orc --create-hcatalog-table --hcatalog-storage-stanza "stored as orcfile"&lt;/STRONG&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 03 May 2016 08:07:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Getting-Error-in-Sqoop-Import-from-Oracle-Mssql-database-in/m-p/132692#M23193</guid>
      <dc:creator>ravi1</dc:creator>
      <dc:date>2016-05-03T08:07:35Z</dc:date>
    </item>
  </channel>
</rss>

