Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here. Want to know more about what has changed? Check out the Community News blog.
I've developed a generic script to sqoop tables from teradata and perform hive import to write to hive tables. Although they work fine, whenever I try to run two sqoop processes at the same time, fetching data from different tables and writing to different hive tables, I'm facing an "AlreadyBeingCreatedException" failure. Sample Error in First Sqoop Error: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException): Failed to create file [/user/<USER>/temp_061649/part-m-00000] for [DFSClient_attempt_1439568235974_1050_m_000000_1_1505507557_1] for client [xx.x.xx.xx], because this file is already being created by [DFSClient_attempt_1439568235974_1049_m_000000_0_-1446753580_1] on [xx.x.xx.xx] Sample Error in Second Sqoop(Running at the same time) INFO mapreduce.Job: Job job_1439568235974_1034 failed with state FAILED due to: Job commit failed: java.io.FileNotFoundException: File hdfs://<SERVER>/user/<USER>/temp_045554/_temporary/1 does not exist. Any help is much appreciated.
... View more