Support Questions

Find answers, ask questions, and share your expertise

Sqoop Export to teradata random results

Hello All

We are trying to export from Hive to teradata using this command

sqoop export -Dorg.apache.sqoop.export.text.dump_data_on_error=true,NullAppender -Dhadoop.root.logger=DEBUG,console -Dsqoop.export.records.per.statement=20 -Dsqoop.export.statements.per.transaction=20 \ --driver com.teradata.jdbc.TeraDriver \ --connect jdbc:teradata:// \ --username dwh_tbda \ --password Teradata_2017 \ --table teratada_table \ --export-dir "hdfs_file" \ --input-null-non-string '\\N' \ --input-null-string '\\N' \ --input-fields-terminated-by '|' \ --num-mappers 20 \ --verbose \ --direct \ -- --output-method internal.fastload

It works ok some time, the results are random and some times export all the records, some times 0 records, or any value beetwen 0 and total. The yarn log is like :

0b1f030 sess=0 connect timed out at Method) at at at at at at$ at org.apache.sqoop.mapreduce.ExportOutputFormat.getRecordWriter( at org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>( at org.apache.hadoop.mapred.MapTask.runNewMapper( at at org.apache.hadoop.mapred.YarnChild$ at Method) at at at org.apache.hadoop.mapred.YarnChild.main( Caused by: java.sql.SQLException: [Teradata JDBC Driver] [TeraJDBC] [Error 1277] [SQLState 08S01] Login timeout for Connection to Wed Apr 11 16:39:10 COT 2018 socket orig= cid=50b1f030 sess=0 connect timed out at Method) at at at at at at$

Some one can suggets a reason to explain whats is happening?

Thank you


Super Collaborator

hi @alvaro andres tovar martinez,

Apparently the issue is with tenacity in teradata, which controls the retry duration for a fastload/mlod operation in teradata.

when you use the fastlad, that plays a major role in busy cluster, the best thing you can do is to ensure that your teradata workload is capable enough to accept the connection when you requested it, no throttles controls your connection at that time or turn off the fastload and do it in normal mode (only if your data is small ) or just keep retry till it get connected in your logic.

more on the teradata jdbc dirver can be found here, but not much help on implementing the tenacity/sleep options as that was not possible when you use JDBC mode.

Hope this helps !!

Thank you for your help.

Y have changed to --output-method batch.insert and run ok, but it takes too many time (all the day).

I am looking for alternatives to load 20 millons of rows. I am trying to create a spark/scala/sqoop script, so that using jdbc and sqoop make a parallel load in multiple tables, and exporting all the partitions of the table. At the side of Teradata i will make a View to consult all the exported tables.

I will try also the fast load for csv files using de jdbc driver instance of sqoop.

Thank you