Member since
11-02-2017
5
Posts
1
Kudos Received
0
Solutions
11-14-2017
02:18 PM
Thanks a lot..It really worked like a charm.. import completes successfully.. But getting a error like (ERROR ATTATCHED) Error: java.lang.RuntimeException: java.lang.RuntimeException: java.sql.SQLRecoverableException: No more data to read from socket Can you put some light on this cause ? 17/11/14 18:41:54 INFO mapreduce.Job: Running job: job_1510681534788_0007
17/11/14 18:42:02 INFO mapreduce.Job: Job job_1510681534788_0007 running in uber mode : false
17/11/14 18:42:02 INFO mapreduce.Job: map 0% reduce 0%
17/11/14 18:42:30 INFO mapreduce.Job: map 20% reduce 0%
17/11/14 18:43:01 INFO mapreduce.Job: map 40% reduce 0%
17/11/14 18:43:02 INFO mapreduce.Job: map 60% reduce 0%
17/11/14 18:43:14 INFO mapreduce.Job: map 80% reduce 0%
17/11/14 18:43:35 INFO mapreduce.Job: Task Id : attempt_1510681534788_0007_m_000003_0, Status : FAILED
Error: java.lang.RuntimeException: java.lang.RuntimeException: java.sql.SQLRecoverableException: No more data to read from socket
at org.apache.sqoop.mapreduce.db.DBInputFormat.setConf(DBInputFormat.java:167)
at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:749)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162)
Caused by: java.lang.RuntimeException: java.sql.SQLRecoverableException: No more data to read from socket
at org.apache.sqoop.mapreduce.db.DBInputFormat.getConnection(DBInputFormat.java:220)
at org.apache.sqoop.mapreduce.db.DBInputFormat.setConf(DBInputFormat.java:165)
... 9 more
Caused by: java.sql.SQLRecoverableException: No more data to read from socket
at oracle.jdbc.driver.SQLStateMapping.newSQLException(SQLStateMapping.java:281)
at oracle.jdbc.driver.DatabaseError.newSQLException(DatabaseError.java:118)
at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:224)
at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:296)
at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:539)
at oracle.jdbc.driver.T4CMAREngine.unmarshalUB1(T4CMAREngine.java:1091)
at oracle.jdbc.driver.T4CMAREngine.unmarshalSB1(T4CMAREngine.java:1040)
at oracle.jdbc.driver.T4CTTIoauthenticate.receiveOauth(T4CTTIoauthenticate.java:814)
at oracle.jdbc.driver.T4CConnection.logon(T4CConnection.java:428)
at oracle.jdbc.driver.PhysicalConnection.<init>(PhysicalConnection.java:494)
... 10 more
17/11/14 18:43:55 INFO mapreduce.Job: map 100% reduce 0%
17/11/14 18:43:55 INFO mapreduce.Job: Job job_1510681534788_0007 completed successfully
17/11/14 18:43:56 INFO mapreduce.Job: Counters: 31
File System Counters
FILE: Number of bytes read=0
FILE: Number of bytes written=750200
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=579
HDFS: Number of bytes written=136
HDFS: Number of read operations=20
HDFS: Number of large read operations=0
HDFS: Number of write operations=10
Job Counters
Failed map tasks=1
Launched map tasks=6
Other local map tasks=6
Total time spent by all maps in occupied slots (ms)=310816
Total time spent by all reduces in occupied slots (ms)=0
Total time spent by all map tasks (ms)=310816
Total vcore-seconds taken by all map tasks=310816
Total megabyte-seconds taken by all map tasks=77704000
Map-Reduce Framework
Map input records=4
Map output records=4
Input split bytes=579
Spilled Records=0
Failed Shuffles=0
Merged Map outputs=0
GC time elapsed (ms)=6122
CPU time spent (ms)=8900
Physical memory (bytes) snapshot=773640192
Virtual memory (bytes) snapshot=4145668096
Total committed heap usage (bytes)=665845760
File Input Format Counters
Bytes Read=0
File Output Format Counters
Bytes Written=136
17/11/14 18:43:56 INFO mapreduce.ImportJobBase: Transferred 136 bytes in 125.5593 seconds (1.0832 bytes/sec)
17/11/14 18:43:56 INFO mapreduce.ImportJobBase: Retrieved 4 records.
17/11/14 18:43:56 INFO util.AppendUtils: Appending to directory orders
17/11/14 18:43:56 INFO util.AppendUtils: Using found partition 10
17/11/14 18:43:56 INFO tool.ImportTool: Incremental import complete! To run another incremental import of all data following this import, supply the following arguments:
17/11/14 18:43:56 INFO tool.ImportTool: --incremental lastmodified
17/11/14 18:43:56 INFO tool.ImportTool: --check-column ORDER_TIMESTAMP
17/11/14 18:43:56 INFO tool.ImportTool: --last-value 2017-11-14 19:29:52.0
17/11/14 18:43:56 INFO tool.ImportTool: (Consider saving this with 'sqoop job --create')
Will be connecting you whenever getting any issue..You are really a champ..Once again thank you dear..
... View more
11-13-2017
08:58 PM
ORA-01840: input value not long enough for date format I have a table named DEMO_ORDERS in oracle like (image attatched) SQOOP INCREMENTAL IMPORT: sqoop import --connect jdbc:oracle:thin:system/system@192.168.XXX.XXX:1521:xe --username sonu -password sonu --table DEMO_ORDERS --target-dir '/user/root/orders/' --as-textfile --split-by CUSTOMER_ID --incremental lastmodified --check-column ORDER_TIMESTAMP --last-value 0 --num-mappers 5 --fields-terminated-by ','; While importing data from Oracle to HDFS using Sqoop i am getting below error.. 17/11/14 01:12:37 INFO manager.OracleManager: Time zone has been set to GMT
17/11/14 01:12:37 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM DEMO_ORDERS t WHERE 1=0
17/11/14 01:12:37 INFO tool.ImportTool: Incremental import based on column ORDER_TIMESTAMP
17/11/14 01:12:37 INFO tool.ImportTool: Lower bound value: TO_TIMESTAMP('0', 'YYYY-MM-DD HH24:MI:SS.FF')
17/11/14 01:12:37 INFO tool.ImportTool: Upper bound value: TO_TIMESTAMP('2017-11-14 01:58:34.0', 'YYYY-MM-DD HH24:MI:SS.FF')
17/11/14 01:12:37 INFO mapreduce.ImportJobBase: Beginning import of DEMO_ORDERS
17/11/14 01:12:37 INFO impl.TimelineClientImpl: Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
17/11/14 01:12:38 INFO client.RMProxy: Connecting to ResourceManager at sandbox.hortonworks.com/192.168.144.128:8050
17/11/14 01:12:39 INFO db.DBInputFormat: Using read commited transaction isolation
17/11/14 01:12:39 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT MIN(CUSTOMER_ID), MAX(CUSTOMER_ID) FROM DEMO_ORDERS WHERE ( ORDER_TIMESTAMP >= TO_TIMESTAMP('0', 'YYYY-MM-DD HH24:MI:SS.FF') AND ORDER_TIMESTAMP < TO_TIMESTAMP('2017-11-14 01:58:34.0', 'YYYY-MM-DD HH24:MI:SS.FF') )
17/11/14 01:12:39 INFO mapreduce.JobSubmitter: Cleaning up the staging area /user/root/.staging/job_1510613991307_0011
17/11/14 01:12:39 ERROR tool.ImportTool: Encountered IOException running import job: java.io.IOException: java.sql.SQLDataException: ORA-01840: input value not long enough for date format HELP ME TO COME OUT OF THIS..PLZ @Sindhu @Shu @Shivpriya Tamboskar
... View more
Labels:
- Labels:
-
Apache Sqoop
11-03-2017
02:36 PM
1 Kudo
@Shu Thank You very much for your brief explanation making it clear.. Actually i am working on a dummy table without timestamp. Hope this will help me with timestamp column.
... View more
11-02-2017
08:05 PM
Step 1: I Sqoop imported data into a HDFS location from Oracle in Avro format using following command. sqoop import --connect jdbc:oracle:thin:system/system@192.XXX.56.188:1521:xe --username xxx -password xxx --table DOB --target-dir '/user/root/DOB_AVRO/' --as-avrodatafile --num-mappers 1; Step 2: Then Created a External table on top of that data using following command. create external table avrodata_new > ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' > STORED as INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' > OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' > location '/user/root/DOB_AVRO/' > TBLPROPERTIES ('avro.schema.url'='hdfs://sandbox.hortonworks.com/user/root/DOB.avsc'); Step 3: hive> select * from avrodata_new; OK SOURYA 477360000000 PATTANAYAK 1 SONU 507945600000 RANA 2 SAROJ NULL BEHERA NULL MANAS NULL NULL 3 TANKA 623980800000 GIRI 4 Time taken: 1.383 seconds, Fetched: 5 row(s) Step 4: Created another table orctab with schema in ORC format using following command. hive> create external table orctab (name string,s_date date,l_name string,id int) > row format delimited > fields terminated by '\t' > lines terminated by '\n' > stored as orc > location '/user/root/orctable'; OK
Time taken: 1.868 seconds Step 5: Inserted the data from avrodata_new to orctab. hive> insert into orctab select name,to_date(timestamp(s_date)),l_name,cast(id as Int) from avrodata_new; Step 6: hive> describe orctab; OK name string s_date date l_name string id int Time taken: 0.675 seconds, Fetched: 4 row(s) hive> select * from orctab; OK SOURYA 1985-02-16 PATTANAYAK 1 SONU 1986-02-05 RANA 2 SAROJ NULL BEHERA NULL MANAS NULL NULL 3 TANKA 1989-10-10 GIRI 4 Time taken: 1.388 seconds, Fetched: 5 row(s) QUESTION: Now i have added 2 records to that HDFS location ('/user/root/DOB_AVRO/') or to avrodata_new table. How can i import only those 2 new records from avrodata_new table to orctab table.
... View more
Labels:
- Labels:
-
Apache Hive