Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Sqoop job is being KILLED when executed though hue

Sqoop job is being KILLED when executed though hue

Explorer

Hello,

 

When I try to run a Sqoop import command through Hue, the job is KILLED always, and I get the following errors in the log:

 

Caused by: java.lang.SecurityException: Intercepted System.exit(1)
	at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)

then after that:

 

Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]

any solution and explanation why this is happening? 

15 REPLIES 15

Re: Sqoop job is being KILLED when executed though hue

Guru
Hi,

Can you please share the content of workflow.xml and job.properties files here for me to have a look? Do you use "--query" options in Sqoop?

Also, what version of CDH are you using?

Cheers
Eric

Re: Sqoop job is being KILLED when executed though hue

Explorer

Sorry, but where I can find the workflow.xml and the job.properties files?

 

And the following is the Sqoop import command I am trying to execute:

sqoop import \
--connect 'jdbc:sqlserver://myURL;database=myDB' \
--username user --password pass \
--table BigDataTest -m 1 --check-column lastmodified \
--merge-key id \
--incremental lastmodified \
--compression-codec=snappy \
--target-dir /user/hive/warehouse/dwh_db.db/bigdatatest \
--hive-table bigDataTest \
--map-column-java lastmodified=String  \
--class-name BigDataTest \
--fields-terminated-by '\001'  --fields-terminated-by '\n'

Re: Sqoop job is being KILLED when executed though hue

Explorer

Hello Eric .... any solution for this?!

 

I have version CHD 6.2
And I am trying to run it from Hue

Re: Sqoop job is being KILLED when executed though hue

Guru
Hi,

Since you are not using "--query" option, can you share the full error from the launcher log for review?

Just trying to get a bigger picture on the issue by checking what happened before and after.

Cheers
Eric
Highlighted

Re: Sqoop job is being KILLED when executed though hue

Explorer

That is a very long log....

This forum does not allow more than 50K char.s ....
The following is the last 50K char.s of the log generated:

 

dfs.namenode.checkpoint.dir : file://${hadoop.tmp.dir}/dfs/namesecondary
 dfs.webhdfs.rest-csrf.browser-useragents-regex : ^Mozilla.*,^Opera.*
 dfs.namenode.top.windows.minutes : 1,5,25
 dfs.client.use.legacy.blockreader.local : false
 mapreduce.job.maxtaskfailures.per.tracker : 3
 mapreduce.shuffle.max.connections : 0
 net.topology.node.switch.mapping.impl : org.apache.hadoop.net.ScriptBasedMapping
 hadoop.kerberos.keytab.login.autorenewal.enabled : false
 yarn.client.application-client-protocol.poll-interval-ms : 200
 mapreduce.fileoutputcommitter.marksuccessfuljobs : true
 yarn.nodemanager.localizer.address : ${yarn.nodemanager.hostname}:8040
 dfs.namenode.list.cache.pools.num.responses : 100
 nfs.server.port : 2049
 dfs.namenode.https-address.NameServiceOne.namenode417 : masternode:9871
 hadoop.proxyuser.HTTP.hosts : *
 dfs.checksum.type : CRC32C
 fs.s3a.readahead.range : 64K
 dfs.client.read.short.circuit.replica.stale.threshold.ms : 1800000
 dfs.ha.namenodes.NameServiceOne : namenode417,namenode434
 ha.zookeeper.parent-znode : /hadoop-ha
 yarn.sharedcache.admin.thread-count : 1
 yarn.nodemanager.resource.cpu-vcores : -1
 mapreduce.jobhistory.http.policy : HTTP_ONLY
 fs.s3a.attempts.maximum : 20
 dfs.datanode.lazywriter.interval.sec : 60
 yarn.log-aggregation.retain-check-interval-seconds : -1
 yarn.resourcemanager.node-ip-cache.expiry-interval-secs : -1
 yarn.timeline-service.client.fd-clean-interval-secs : 60
 fs.wasbs.impl : org.apache.hadoop.fs.azure.NativeAzureFileSystem$Secure
 dfs.federation.router.reader.count : 1
 hadoop.ssl.keystores.factory.class : org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory
 hadoop.zk.num-retries : 1000
 mapreduce.job.split.metainfo.maxsize : 10000000
 hadoop.security.random.device.file.path : /dev/urandom
 yarn.client.nodemanager-connect.max-wait-ms : 180000
 yarn.app.mapreduce.client-am.ipc.max-retries : 3
 dfs.namenode.snapshotdiff.allow.snap-root-descendant : true
 yarn.nodemanager.container-diagnostics-maximum-size : 10000
 yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage : false
 dfs.namenode.ec.system.default.policy : RS-6-3-1024k
 dfs.replication.max : 512
 dfs.datanode.https.address : 0.0.0.0:9865
 dfs.ha.standby.checkpoints : true
 ipc.client.kill.max : 10
 mapreduce.job.committer.setup.cleanup.needed : true
 dfs.client.domain.socket.data.traffic : false
 yarn.nodemanager.localizer.cache.target-size-mb : 10240
 yarn.resourcemanager.admin.client.thread-count : 1
 hadoop.security.group.mapping.ldap.connection.timeout.ms : 60000
 yarn.timeline-service.store-class : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore
 yarn.resourcemanager.nm-container-queuing.queue-limit-stdev : 1.0f
 yarn.resourcemanager.zk-appid-node.split-index : 0
 hadoop.tmp.dir : /tmp/hadoop-${user.name}
 dfs.domain.socket.disable.interval.seconds : 1
 fs.s3a.etag.checksum.enabled : false
 hadoop.security.kms.client.failover.sleep.base.millis : 100
 yarn.node-labels.configuration-type : centralized
 fs.s3a.retry.interval : 500ms
 dfs.datanode.http.internal-proxy.port : 0
 yarn.timeline-service.ttl-ms : 604800000
 mapreduce.task.exit.timeout.check-interval-ms : 20000
 oozie.sqoop.args.7 : \
--table
 oozie.sqoop.args.8 : category
 mapreduce.map.speculative : false
 oozie.sqoop.args.5 : --password
 oozie.sqoop.args.6 : myUsername
 yarn.nodemanager.linux-container-executor.cgroups.delete-timeout-ms : 1000
 yarn.timeline-service.recovery.enabled : false
 oozie.sqoop.args.9 : -m
 yarn.nodemanager.recovery.dir : ${hadoop.tmp.dir}/yarn-nm-recovery
 mapreduce.job.counters.max : 120
 dfs.namenode.name.cache.threshold : 10
 oozie.sqoop.args.0 : import
 dfs.namenode.caching.enabled : true
 dfs.namenode.max.full.block.report.leases : 6
 oozie.sqoop.args.3 : \
--username
 yarn.nodemanager.linux-container-executor.cgroups.delete-delay-ms : 20
 dfs.namenode.max.extra.edits.segments.retained : 10000
 oozie.sqoop.args.4 : myUsername
 dfs.webhdfs.user.provider.user.pattern : ^[A-Za-z_][A-Za-z0-9._-]*[$]?$
 yarn.webapp.ui2.enable : false
 oozie.sqoop.args.1 : \
--connect
 oozie.sqoop.args.2 : 'jdbc:sqlserver://myServer;database=myDB'
 dfs.client.mmap.enabled : true
 mapreduce.map.log.level : INFO
 dfs.datanode.ec.reconstruction.threads : 8
 hadoop.fuse.timer.period : 5
 yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms : 1000
 hadoop.zk.timeout-ms : 10000
 ha.health-monitor.check-interval.ms : 1000
 dfs.client.hedged.read.threshold.millis : 500
 yarn.resourcemanager.fs.state-store.retry-interval-ms : 1000
 mapreduce.output.fileoutputformat.compress : false
 yarn.sharedcache.store.in-memory.staleness-period-mins : 10080
 dfs.client.write.byte-array-manager.count-limit : 2048
 mapreduce.application.framework.path : hdfs://NameServiceOne//user/yarn/mapreduce/mr-framework/3.0.0-cdh6.2.0-mr-framework.tar.gz#mr-framework
 hadoop.security.group.mapping.providers.combined : true
 fs.AbstractFileSystem.har.impl : org.apache.hadoop.fs.HarFs
 mapreduce.job.running.map.limit : 0
 yarn.nodemanager.webapp.address : ${yarn.nodemanager.hostname}:8042
 mapreduce.reduce.input.buffer.percent : 0.0
 mapreduce.job.cache.files : hdfs://NameServiceOne/user/hue/oozie/deployments/_admin_-oozie-312-1560674439.04/lib/hive-site.xml#hive-site.xml,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-exec-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-security-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slider-core-0.90.2-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-repackaged-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcodings-1.0.18.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fst-2.50.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ST4-4.0.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-protobuf.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-0.23.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ecj-4.4.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libthrift-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-crypto-1.0.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-schemas-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/transaction-api-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-webapp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joda-time-2.9.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-core-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-hcatalog-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-tez.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-commons-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-container-servlet-core-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-java7-2.4.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-utils-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-hive.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stringtemplate-3.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-core-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bonecp-0.8.0.RELEASE.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-core-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/htrace-core4-4.1.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-api-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-client-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-encoding.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-ipc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-cli.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fastutil-7.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guava-11.0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oro-2.0.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-api-2.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-tree-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcpkix-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-pool-1.5.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jpam-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/osgi-resource-locator-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-annotations-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpcore-4.4.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-archives.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slf4j-api-1.7.25.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-hbase-compat-1.0-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hsqldb-1.8.0.10.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-resourcemanager.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jaas-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-linq4j-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-codec-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/gson-2.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol-shaded.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-api-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jta-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-dbcp-1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-common-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aggdesigner-algorithm-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-json-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/opencsv-2.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-column.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.ws.rs-api-2.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-locator-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javolution-5.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-web-proxy.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-ant.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/groovy-all-2.4.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop-bundle.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-http-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-api-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/mssql-jdbc-6.2.1.jre7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-common-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joni-2.1.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-2.7.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jstl-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jndi-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libfb303-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-java-1.1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.el-3.0.1-b11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-plus-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-httpclient-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-netty.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/re2j-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-io-2.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javassist-3.20.0-GA.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-media-jaxb-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-curator-2.12.0.pom#apache-curator-2.12.0.pom,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-classification.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-databind-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-replication.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-storage-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet-api-3.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-client-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.annotation-api-1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-framework-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-scheduler.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcprov-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-procedure.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-rewrite-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-serde.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-spec-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-druid-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.jdo-3.2.0-m3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-core-4.1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-jvm-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/logredactor-2.0.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-web-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jsp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/java-util-1.9.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcommander-1.30.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-metastore.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-runner-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-2.6.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compress-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-assistedinject-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-slf4j-impl-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ehcache-3.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop2-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpclient-4.5.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/objenesis-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-common-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-server-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/velocity-1.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/janino-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-http.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stax-api-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang3-3.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avatica-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-runtime-3.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-registry.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/leveldbjni-all-1.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-util-ajax-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jdo-api-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jline-2.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-impl-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-asl-1.9.13.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-io-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-launcher-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/audience-annotations-0.5.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-hadoop-compatibility.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-2.3.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/derby-10.14.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/findbugs-annotations-1.3.9-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/xz-1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/geronimo-jcache_1.0_spec-1.0-alpha-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-api-jdo-4.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/netty-3.10.6.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-xml-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ivy-2.4.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/paranamer-2.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compiler-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-miscellaneous.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-mapred-hadoop2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-applicationhistoryservice.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-rdbms-4.1.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-core-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service-rpc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jsr305-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-mapper-asl-1.9.13-cloudera.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-20090211.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/validation-api-1.1.0.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-guava-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-zookeeper-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-format.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2.17.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/dropwizard-metrics-hadoop-metrics2-reporter-0.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-annotations-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-io-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-jackson.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/disruptor-3.3.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-orc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie.jar
 dfs.client.slow.io.warning.threshold.ms : 30000
 fs.s3a.multipart.size : 100M
 yarn.app.mapreduce.am.job.committer.commit-window : 10000
 dfs.qjournal.new-epoch.timeout.ms : 120000
 yarn.timeline-service.webapp.rest-csrf.enabled : false
 hadoop.proxyuser.flume.hosts : *
 dfs.namenode.edits.asynclogging : true
 yarn.timeline-service.reader.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineReaderImpl
 yarn.app.mapreduce.am.staging-dir.erasurecoding.enabled : false
 dfs.blockreport.incremental.intervalMsec : 0
 dfs.datanode.network.counts.cache.max.size : 2147483647
 dfs.namenode.https-address.NameServiceOne.namenode434 : node3:9871
 yarn.timeline-service.writer.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineWriterImpl
 mapreduce.ifile.readahead : true
 dfs.qjournal.get-journal-state.timeout.ms : 120000
 yarn.timeline-service.entity-group-fs-store.summary-store : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore
 dfs.client.socketcache.capacity : 16
 fs.s3a.s3guard.ddb.table.create : false
 dfs.client.retry.policy.spec : 10000,6,60000,10
 mapreduce.output.fileoutputformat.compress.codec : org.apache.hadoop.io.compress.DefaultCodec
 fs.s3a.socket.recv.buffer : 8192
 dfs.datanode.fsdatasetcache.max.threads.per.volume : 4
 dfs.namenode.reencrypt.batch.size : 1000
 yarn.sharedcache.store.in-memory.initial-delay-mins : 10
 mapreduce.jobhistory.webapp.address : masternode:19888
 fs.adl.impl : org.apache.hadoop.fs.adl.AdlFileSystem
 fs.AbstractFileSystem.gs.impl : com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS
 mapreduce.task.userlog.limit.kb : 0
 fs.s3a.connection.ssl.enabled : true
 yarn.router.rmadmin.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.rmadmin.DefaultRMAdminRequestInterceptor
 yarn.sharedcache.webapp.address : 0.0.0.0:8788
 hadoop.fuse.connection.timeout : 300
 dfs.http.client.retry.policy.spec : 10000,6,60000,10
 yarn.resourcemanager.rm.container-allocation.expiry-interval-ms : 600000
 ipc.server.max.connections : 0
 yarn.app.mapreduce.am.resource.mb : 3072
 hadoop.security.groups.cache.secs : 300
 dfs.datanode.peer.stats.enabled : false
 dfs.replication : 3
 mapreduce.shuffle.transfer.buffer.size : 131072
 dfs.namenode.audit.log.async : false
 hadoop.security.group.mapping.ldap.directory.search.timeout : 10000
 dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold : 10737418240
 dfs.datanode.disk.check.timeout : 10m
 yarn.app.mapreduce.client-am.ipc.max-retries-on-timeouts : 3
 fs.s3a.committer.threads : 8
 dfs.checksum.combine.mode : MD5MD5CRC
 yarn.resourcemanager.nodemanager-graceful-decommission-timeout-secs : 3600
 yarn.scheduler.maximum-allocation-vcores : 6
 yarn.nodemanager.sleep-delay-before-sigkill.ms : 250
 fs.AbstractFileSystem.abfs.impl : org.apache.hadoop.fs.azurebfs.Abfs
 mapreduce.job.acl-modify-job :  
 fs.automatic.close : true
 fs.azure.sas.expiry.period : 90d
 dfs.qjm.operations.timeout : 60s
 hadoop.proxyuser.httpfs.hosts : *
 dfs.namenode.stale.datanode.minimum.interval : 3
 dfs.namenode.reencrypt.edek.threads : 10
 dfs.federation.router.store.membership.expiration : 300000
 hadoop.security.groups.cache.background.reload.threads : 3
 mapreduce.input.fileinputformat.list-status.num-threads : 1
 hadoop.security.group.mapping.ldap.posix.attr.gid.name : gidNumber
 dfs.namenode.acls.enabled : false
 dfs.client.short.circuit.replica.stale.threshold.ms : 1800000
 dfs.namenode.resource.du.reserved : 104857600
 dfs.federation.router.connection.clean.ms : 10000
 dfs.client.server-defaults.validity.period.ms : 3600000
 dfs.federation.router.metrics.class : org.apache.hadoop.hdfs.server.federation.metrics.FederationRPCPerformanceMonitor
 mapreduce.shuffle.listen.queue.size : 128
 mapreduce.jobhistory.intermediate-done-dir : ${yarn.app.mapreduce.am.staging-dir}/history/done_intermediate
 mapreduce.client.libjars.wildcard : true
 dfs.federation.router.cache.ttl : 60000
 yarn.nodemanager.recovery.compaction-interval-secs : 3600
 dfs.namenode.edits.noeditlogchannelflush : false
 mapreduce.reduce.shuffle.input.buffer.percent : 0.70
 yarn.http.policy : HTTP_ONLY
 mapreduce.map.maxattempts : 4
 dfs.namenode.audit.loggers : default
 io.serializations : org.apache.hadoop.io.serializer.WritableSerialization, org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization, org.apache.hadoop.io.serializer.avro.AvroReflectSerialization
 hadoop.security.groups.cache.warn.after.ms : 5000
 dfs.client.write.byte-array-manager.count-reset-time-period-ms : 10000
 yarn.nodemanager.webapp.rest-csrf.custom-header : X-XSRF-Header
 yarn.app.mapreduce.am.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH
 dfs.namenode.snapshot.capture.openfiles : true
 yarn.node-labels.fs-store.impl.class : org.apache.hadoop.yarn.nodelabels.FileSystemNodeLabelsStore
 hadoop.http.cross-origin.allowed-methods : GET,POST,HEAD
 dfs.qjournal.queued-edits.limit.mb : 10
 mapreduce.jobhistory.webapp.rest-csrf.enabled : false
 dfs.http.policy : HTTP_ONLY
 dfs.balancer.max-size-to-move : 10737418240
 dfs.datanode.sync.behind.writes.in.background : false
 hadoop.zk.acl : world:anyone:rwcda
 yarn.nodemanager.container.stderr.pattern : {*stderr*,*STDERR*}
 dfs.namenode.reencrypt.throttle.limit.updater.ratio : 1.0
 mapreduce.cluster.local.dir : ${hadoop.tmp.dir}/mapred/local
 hadoop.kerberos.kinit.command : kinit
 dfs.namenode.secondary.https-address : 0.0.0.0:9869
 dfs.namenode.metrics.logger.period.seconds : 600
 dfs.block.access.token.lifetime : 600
 dfs.ha.automatic-failover.enabled.NameServiceOne : true
 dfs.namenode.delegation.token.max-lifetime : 604800000
 dfs.datanode.drop.cache.behind.writes : false
 dfs.mover.address : 0.0.0.0:0
 dfs.block.placement.ec.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyRackFaultTolerant
 dfs.namenode.num.extra.edits.retained : 1000000
 ipc.client.connect.max.retries.on.timeouts : 45
 fs.viewfs.rename.strategy : SAME_MOUNTPOINT
 fs.client.resolve.topology.enabled : false
 hadoop.proxyuser.hive.hosts : *
 yarn.resourcemanager.node-labels.provider.fetch-interval-ms : 1800000
 yarn.nodemanager.container-metrics.enable : true
 mapreduce.job.map.output.collector.class : org.apache.hadoop.mapred.MapTask$MapOutputBuffer
 fs.s3a.fast.upload.buffer : disk
 ha.health-monitor.connect-retry-interval.ms : 1000
 dfs.namenode.edekcacheloader.initial.delay.ms : 3000
 dfs.edit.log.transfer.bandwidthPerSec : 0
 dfs.ha.tail-edits.in-progress : false
 dfs.federation.router.heartbeat.interval : 5000
 ssl.client.truststore.reload.interval : 10000
 dfs.client.datanode-restart.timeout : 30s
 io.mapfile.bloom.size : 1048576
 hadoop.security.kms.client.authentication.retry-count : 1
 dfs.client-write-packet-size : 65536
 fs.ftp.data.connection.mode : ACTIVE_LOCAL_DATA_CONNECTION_MODE
 fs.swift.impl : org.apache.hadoop.fs.swift.snative.SwiftNativeFileSystem
 yarn.resourcemanager.webapp.rest-csrf.methods-to-ignore : GET,OPTIONS,HEAD
 mapreduce.job.max.map : -1
 yarn.app.mapreduce.shuffle.log.backups : 0
 ftp.blocksize : 67108864
 dfs.namenode.kerberos.principal.pattern : *
 yarn.resourcemanager.scheduler.monitor.enable : false
 dfs.webhdfs.socket.connect-timeout : 60s
 dfs.namenode.replication.max-streams : 2
 nfs.allow.insecure.ports : true
 yarn.sharedcache.nm.uploader.thread-count : 20
 dfs.federation.router.admin.enable : true
 yarn.app.mapreduce.client.job.retry-interval : 2000
 yarn.scheduler.configuration.store.max-logs : 1000
 hadoop.security.authorization : false
 yarn.timeline-service.version : 1.0f
 yarn.am.liveness-monitor.expiry-interval-ms : 600000
 fs.har.impl.disable.cache : true
 hadoop.proxyuser.hdfs.hosts : *
 mapreduce.job.reduce.slowstart.completedmaps : 0.8
 yarn.timeline-service.leveldb-timeline-store.path : ${hadoop.tmp.dir}/yarn/timeline
 dfs.namenode.upgrade.domain.factor : ${dfs.replication}
 mapreduce.jobhistory.minicluster.fixed.ports : false
 mapreduce.application.classpath : $HADOOP_CLIENT_CONF_DIR,$PWD/mr-framework/*,$MR2_CLASSPATH
 yarn.resourcemanager.delegation.token.max-lifetime : 604800000
 yarn.resourcemanager.ha.automatic-failover.enabled : true
 mapreduce.reduce.java.opts : -Djava.net.preferIPv4Stack=true
 dfs.datanode.socket.write.timeout : 480000
 dfs.namenode.accesstime.precision : 3600000
 dfs.namenode.redundancy.considerLoad.factor : 2.0
 yarn.resourcemanager.store.class : org.apache.hadoop.yarn.server.resourcemanager.recovery.FileSystemRMStateStore
 io.mapfile.bloom.error.rate : 0.005
 yarn.nodemanager.webapp.rest-csrf.enabled : false
 yarn.timeline-service.leveldb-state-store.path : ${hadoop.tmp.dir}/yarn/timeline
 hadoop.proxyuser.hive.groups : *
 dfs.federation.router.rpc-address : 0.0.0.0:8888
 fs.s3a.committer.staging.unique-filenames : true
 dfs.namenode.support.allow.format : true
 yarn.scheduler.configuration.zk-store.parent-path : /confstore
 dfs.content-summary.limit : 5000
 yarn.timeline-service.writer.flush-interval-seconds : 60
 yarn.nodemanager.container-executor.class : org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor
 dfs.namenode.posix.acl.inheritance.enabled : true
 dfs.datanode.outliers.report.interval : 30m
 hadoop.security.kms.client.encrypted.key.cache.low-watermark : 0.3f
 dfs.namenode.top.enabled : true
 yarn.app.mapreduce.shuffle.log.separate : true
 hadoop.user.group.static.mapping.overrides : dr.who=;
 dfs.federation.router.http-address : 0.0.0.0:50071
 fs.s3a.retry.throttle.interval : 1000ms
 yarn.nodemanager.amrmproxy.address : 0.0.0.0:8049
 mapreduce.jobhistory.webapp.rest-csrf.custom-header : X-XSRF-Header
 yarn.webapp.xfs-filter.enabled : true
 dfs.client.cached.conn.retry : 3
 dfs.client.key.provider.cache.expiry : 864000000
 dfs.namenode.path.based.cache.refresh.interval.ms : 30000
 yarn.nodemanager.collector-service.thread-count : 5
 dfs.block.replicator.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyDefault
 dfs.namenode.fs-limits.max-directory-items : 1048576
 dfs.ha.log-roll.period : 120s
 yarn.nodemanager.runtime.linux.docker.capabilities : CHOWN,DAC_OVERRIDE,FSETID,FOWNER,MKNOD,NET_RAW,SETGID,SETUID,SETFCAP,SETPCAP,NET_BIND_SERVICE,SYS_CHROOT,KILL,AUDIT_WRITE
 yarn.nodemanager.distributed-scheduling.enabled : false
 ipc.client.fallback-to-simple-auth-allowed : false
 yarn.minicluster.fixed.ports : false
 yarn.nodemanager.remote-app-log-dir : /tmp/logs
 yarn.timeline-service.entity-group-fs-store.scan-interval-seconds : 60
 dfs.xframe.enabled : true
 yarn.nodemanager.resource.percentage-physical-cpu-limit : 100
 mapreduce.job.tags : oozie-564a124254f1fd53cb03553181f7e603
 dfs.namenode.fs-limits.max-xattr-size : 16384
 dfs.datanode.http.address : 0.0.0.0:9864
 dfs.namenode.blocks.per.postponedblocks.rescan : 10000
 fs.s3a.s3guard.cli.prune.age : 86400000
 dfs.web.authentication.filter : org.apache.hadoop.hdfs.web.AuthFilter
 dfs.namenode.maintenance.replication.min : 1
 hadoop.jetty.logs.serve.aliases : true
 dfs.webhdfs.ugi.expire.after.access : 600000
 dfs.namenode.max.op.size : 52428800
 mapreduce.jobhistory.admin.acl : *
 mapreduce.job.reducer.unconditional-preempt.delay.sec : 300
 yarn.app.mapreduce.am.hard-kill-timeout-ms : 10000
 yarn.resourcemanager.display.per-user-apps : false
 yarn.resourcemanager.node-removal-untracked.timeout-ms : 60000
 yarn.resourcemanager.webapp.address : masternode:8088
 mapreduce.jobhistory.recovery.enable : false
 yarn.sharedcache.store.in-memory.check-period-mins : 720
 dfs.client.test.drop.namenode.response.number : 0
 fs.df.interval : 60000
 fs.s3a.assumed.role.session.duration : 30m
 mapreduce.job.cache.limit.max-single-resource-mb : 0
 yarn.timeline-service.enabled : false
 dfs.disk.balancer.block.tolerance.percent : 10
 dfs.webhdfs.netty.high.watermark : 65535
 mapreduce.task.profile : false
 hadoop.http.cross-origin.allowed-headers : X-Requested-With,Content-Type,Accept,Origin
 yarn.router.webapp.address : 0.0.0.0:8089
 dfs.datanode.balance.max.concurrent.moves : 50
 yarn.nodemanager.hostname : 0.0.0.0
 mapreduce.task.exit.timeout : 60000
 yarn.resourcemanager.nm-container-queuing.max-queue-length : 15
 mapreduce.job.token.tracking.ids.enabled : false
 yarn.scheduler.increment-allocation-mb : 512
 fs.s3a.assumed.role.credentials.provider : org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider
 fs.azure.authorization.caching.enable : true
 hadoop.security.kms.client.failover.sleep.max.millis : 2000
 dfs.client.mmap.retry.timeout.ms : 300000
 yarn.resourcemanager.webapp.rest-csrf.custom-header : X-XSRF-Header
 yarn.resourcemanager.nm-container-queuing.max-queue-wait-time-ms : 100
 mapreduce.jobhistory.move.thread-count : 3
 dfs.permissions.enabled : true
 fs.AbstractFileSystem.hdfs.impl : org.apache.hadoop.fs.Hdfs
 yarn.nodemanager.container-localizer.log.level : INFO
 hadoop.http.filter.initializers : org.apache.hadoop.http.lib.StaticUserWebFilter
 yarn.timeline-service.http-authentication.simple.anonymous.allowed : true
 yarn.nodemanager.runtime.linux.docker.allowed-container-networks : host,none,bridge
 dfs.qjournal.accept-recovery.timeout.ms : 120000
 yarn.sharedcache.client-server.thread-count : 50
 fs.s3a.s3guard.ddb.max.retries : 9
 fs.s3a.committer.magic.enabled : false
 yarn.resourcemanager.resource-tracker.address : masternode:8031
 mapreduce.jobhistory.jobname.limit : 50
 dfs.domain.socket.path : /var/run/hdfs-sockets/dn
 dfs.namenode.decommission.blocks.per.interval : 500000
 dfs.qjournal.write-txns.timeout.ms : 20000
 rpc.metrics.quantile.enable : false
 yarn.federation.subcluster-resolver.class : org.apache.hadoop.yarn.server.federation.resolver.DefaultSubClusterResolverImpl
 dfs.namenode.read-lock-reporting-threshold-ms : 5000
 mapreduce.task.timeout : 600000
 yarn.nodemanager.resource.memory-mb : -1
 dfs.datanode.failed.volumes.tolerated : 0
 yarn.nodemanager.disk-health-checker.min-healthy-disks : 0.25
 mapreduce.framework.name : yarn
 mapreduce.fileoutputcommitter.algorithm.version : 2
 yarn.router.clientrm.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.clientrm.DefaultClientRequestInterceptor
 yarn.sharedcache.nested-level : 3
 fs.s3a.connection.timeout : 200000
 hadoop.caller.context.signature.max.size : 40
 hadoop.security.dns.log-slow-lookups.enabled : false
 mapreduce.jobhistory.webapp.https.address : masternode:19890
 file.client-write-packet-size : 65536
 fs.s3a.s3guard.ddb.table.capacity.read : 500
 ipc.client.ping : true
 hadoop.proxyuser.oozie.hosts : *
 yarn.resourcemanager.delayed.delegation-token.removal-interval-ms : 30000
 dfs.client.failover.max.attempts : 15
 dfs.balancer.max-no-move-interval : 60000
 yarn.nodemanager.opportunistic-containers-use-pause-for-preemption : false
 yarn.nodemanager.webapp.cross-origin.enabled : false
 mapreduce.job.encrypted-intermediate-data : false
 dfs.client.read.shortcircuit.streams.cache.expiry.ms : 300000
 yarn.minicluster.control-resource-monitoring : false
 dfs.disk.balancer.enabled : false
 dfs.webhdfs.oauth2.enabled : false
 yarn.nodemanager.health-checker.script.timeout-ms : 1200000
 yarn.resourcemanager.fs.state-store.num-retries : 0
 hadoop.ssl.require.client.cert : false
 mapreduce.jobhistory.keytab : /etc/security/keytab/jhs.service.keytab
 hadoop.security.uid.cache.secs : 14400
 yarn.resourcemanager.ha.automatic-failover.zk-base-path : /yarn-leader-election
 yarn.intermediate-data-encryption.enable : false
 mapreduce.job.speculative.speculative-cap-running-tasks : 0.1
 dfs.datanode.block.id.layout.upgrade.threads : 12
 dfs.client.context : default
 yarn.system-metrics-publisher.enabled : false
 dfs.namenode.delegation.token.renew-interval : 86400000
 yarn.timeline-service.entity-group-fs-store.app-cache-size : 10
 fs.AbstractFileSystem.s3a.impl : org.apache.hadoop.fs.s3a.S3A
 mapreduce.job.redacted-properties : fs.s3a.access.key,fs.s3a.secret.key,fs.adl.oauth2.credential,dfs.adls.oauth2.credential,fs.azure.account.oauth2.client.secret
 yarn.client.load.resource-types.from-server : false
 ipc.client.tcpnodelay : true
 hadoop.proxyuser.httpfs.groups : *
 yarn.resourcemanager.metrics.runtime.buckets : 60,300,1440
 dfs.blockreport.intervalMsec : 21600000
 dfs.datanode.oob.timeout-ms : 1500,0,0,0
 yarn.client.application-client-protocol.poll-timeout-ms : -1
 zlib.compress.level : DEFAULT_COMPRESSION
 mapreduce.job.sharedcache.mode : disabled
 io.map.index.skip : 0
 mapreduce.job.hdfs-servers : ${fs.defaultFS}
 mapreduce.map.output.compress : true
 hadoop.security.kms.client.encrypted.key.cache.num.refill.threads : 2
 dfs.namenode.edekcacheloader.interval.ms : 1000
 mapreduce.task.merge.progress.records : 10000
 yarn.nodemanager.aux-services.mapreduce_shuffle.class : org.apache.hadoop.mapred.ShuffleHandler
 dfs.namenode.missing.checkpoint.periods.before.shutdown : 3
 tfile.fs.output.buffer.size : 262144
 dfs.client.failover.connection.retries : 0
 fs.du.interval : 600000
 dfs.edit.log.transfer.timeout : 30000
 dfs.namenode.top.window.num.buckets : 10
 dfs.data.transfer.server.tcpnodelay : true
 hadoop.zk.retry-interval-ms : 1000
 yarn.sharedcache.uploader.server.address : 0.0.0.0:8046
 dfs.http.client.failover.max.attempts : 15
 fs.s3a.socket.send.buffer : 8192
 dfs.client.block.write.locateFollowingBlock.retries : 7
 hadoop.registry.zk.quorum : localhost:2181
 mapreduce.jvm.system-properties-to-log : os.name,os.version,java.home,java.runtime.version,java.vendor,java.version,java.vm.name,java.class.path,java.io.tmpdir,user.dir,user.name
 hadoop.http.cross-origin.allowed-origins : *
 dfs.namenode.enable.retrycache : true
 dfs.datanode.du.reserved : 0
 hadoop.registry.system.acls : sasl:yarn@, sasl:mapred@, sasl:hdfs@
 yarn.nodemanager.resource-plugins.gpu.docker-plugin.nvidia-docker-v1.endpoint : http://localhost:3476/v1.0/docker/cli
 mapreduce.job.encrypted-intermediate-data.buffer.kb : 128
 dfs.data.transfer.client.tcpnodelay : true
 yarn.resourcemanager.webapp.xfs-filter.xframe-options : SAMEORIGIN
 mapreduce.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH
 mapreduce.task.profile.reduce.params : ${mapreduce.task.profile.params}
 mapreduce.reduce.memory.mb : 0
 hadoop.caller.context.enabled : false
 hadoop.http.authentication.kerberos.principal : HTTP/_HOST@LOCALHOST
 yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb : 0
 dfs.qjournal.prepare-recovery.timeout.ms : 120000
 dfs.datanode.transferTo.allowed : true
 oozie.action.rootlogger.log.level : INFO
 hadoop.security.sensitive-config-keys : 
      secret$
      password$
      ssl.keystore.pass$
      fs.s3.*[Ss]ecret.?[Kk]ey
      fs.s3a.*.server-side-encryption.key
      fs.azure.account.key.*
      credential$
      oauth.*secret
      oauth.*password
      oauth.*token
      hadoop.security.sensitive-config-keys
  
 mapreduce.client.completion.pollinterval : 5000
 dfs.namenode.name.dir.restore : false
 dfs.namenode.full.block.report.lease.length.ms : 300000
 dfs.namenode.secondary.http-address : 0.0.0.0:9868
 hadoop.http.logs.enabled : true
 hadoop.security.group.mapping.ldap.read.timeout.ms : 60000
 yarn.resourcemanager.max-log-aggregation-diagnostics-in-memory : 10
 dfs.namenode.delegation.token.always-use : false
 yarn.resourcemanager.webapp.https.address : masternode:8090
 fs.s3a.retry.throttle.limit : ${fs.s3a.attempts.maximum}
 dfs.client.read.striped.threadpool.size : 18
 mapreduce.job.cache.limit.max-resources : 0
 hadoop.proxyuser.HTTP.groups : *
--------------------

Setting up log4j2
log4j2 configuration file created at /yarn/nm/usercache/admin/appcache/application_1560674082717_0001/container_1560674082717_0001_01_000001/sqoop-log4j2.xml
Sqoop command arguments :
             import
             \
--connect
             'jdbc:sqlserver://myServer;database=myDB'
             \
--username
             myUsername
             --password
             ********
             \
--table
             category
             -m
             1
             --check-column
             LastEditOn
             \
--merge-key
             'Reference
             ID'
             \
--incremental
             lastmodified
             \
--compression-codec=snappy
             \
--target-dir
             /user/hive/warehouse/myDB.db/category
             \
--hive-table
             category
             \
--map-column-hive
             LastEditOn=timestamp,CreatedOn=timestamp
             \
--fields-terminated-by
             '\001'
             --fields-terminated-by
             '\n'
Fetching child yarn jobs
tag id : oozie-564a124254f1fd53cb03553181f7e603
No child applications found
=================================================================

>>> Invoking Sqoop command line now >>>


<<< Invocation of Sqoop command completed <<<

No child hadoop job is executed.
java.lang.reflect.InvocationTargetException
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.oozie.action.hadoop.LauncherAM.runActionMain(LauncherAM.java:410)
	at org.apache.oozie.action.hadoop.LauncherAM.access$300(LauncherAM.java:55)
	at org.apache.oozie.action.hadoop.LauncherAM$2.run(LauncherAM.java:223)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
	at org.apache.oozie.action.hadoop.LauncherAM.run(LauncherAM.java:217)
	at org.apache.oozie.action.hadoop.LauncherAM$1.run(LauncherAM.java:153)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
	at org.apache.oozie.action.hadoop.LauncherAM.main(LauncherAM.java:141)
Caused by: java.lang.SecurityException: Intercepted System.exit(1)
	at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
	at java.lang.Runtime.exit(Runtime.java:107)
	at java.lang.System.exit(System.java:971)
	at org.apache.sqoop.Sqoop.main(Sqoop.java:252)
	at org.apache.oozie.action.hadoop.SqoopMain.runSqoopJob(SqoopMain.java:214)
	at org.apache.oozie.action.hadoop.SqoopMain.run(SqoopMain.java:199)
	at org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:104)
	at org.apache.oozie.action.hadoop.SqoopMain.main(SqoopMain.java:51)
	... 16 more
Intercepting System.exit(1)
Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]
Oozie Launcher, uploading action data to HDFS sequence file: hdfs://NameServiceOne/user/admin/oozie-oozi/0000000-190616123600049-oozie-oozi-W/sqoop-c9e7--sqoop/action-data.seq
12:41:09.783 [main] INFO  org.apache.hadoop.io.compress.CodecPool - Got brand-new compressor [.deflate]
Stopping AM
12:41:09.983 [main] INFO  org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl - Waiting for application to be successfully unregistered.
Callback notification attempts left 0
Callback notification trying http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED
Callback notification to http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED succeeded
Callback notification succeeded

 

Re: Sqoop job is being KILLED when executed though hue

Guru
Hi,

Thanks for sharing the output. From the error, looks like that Sqoop did not parse the command correctly. And I noticed below parameter:

--merge-key
'Reference
ID'

Noticed the line break between Reference and ID? I am not sure if it is formatting issue in the forum or it was like this in your Sqoop job setup. Can you please double check on this? And make sure there is no new line between the two words?

Cheers
Eric

Re: Sqoop job is being KILLED when executed though hue

Explorer

I am sorry but there is nothing wrong with the syntax, as if I run it on the terminal it completes successfully.

I have a doubt regarding security, and that is because of the following line in the log 

aused by: java.lang.SecurityException: Intercepted System.exit(1)
	at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)

I would like to note also that this is the first try to run sqoop script on Hue after a fresh installation of cdh 6.2... so I am afraid there is something that I've missed in the configuration, but I really can't find it :( 

Re: Sqoop job is being KILLED when executed though hue

Guru
Can you please share the workflow.xml contents for review? The new line "Reference" does not seem correct, so I want to see how it is setup in the XML file .

Please also share the screenshot of the Sqoop job configuration in Hue for review.

Cheers
Eric

Re: Sqoop job is being KILLED when executed though hue

Explorer

I hope this is what you want for the workflow:

<workflow-app name="Batch job for query-sqoop1" xmlns="uri:oozie:workflow:0.5">
    <start to="sqoop-fde5"/>
    <kill name="Kill">
        <message>Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
    </kill>
    <action name="sqoop-fde5">
        <sqoop xmlns="uri:oozie:sqoop-action:0.2">
            <job-tracker>${jobTracker}</job-tracker>
            <name-node>${nameNode}</name-node>
            <command>import \
--connect &#39;jdbc:sqlserver://11.11.11.11;database=DBXYZ&#39; \
--username theUser --password thePassword \
--table category -m 1  --check-column LastEditOn \
--merge-key &#39;Reference ID&#39; \
--incremental lastmodified \
--compression-codec=snappy \
--target-dir /user/hive/warehouse/dwh_db_atlas_jrtf.db/category \
--hive-table category \
--map-column-hive LastEditOn=timestamp,CreatedOn=timestamp \
--fields-terminated-by &#39;\001&#39;  --fields-terminated-by &#39;\n&#39;</command>
        </sqoop>
        <ok to="End"/>
        <error to="Kill"/>
    </action>
    <end name="End"/>
</workflow-app>

For the job configuration, I am really not sure where to find it. The one I can reach is a lot of scrolls that can't be taken screenshot in anyway.... so would you please give me the path to the job configuration ?