Created on 05-27-2019 04:03 AM - edited 09-16-2022 07:24 AM
Hello,
When I try to run a Sqoop import command through Hue, the job is KILLED always, and I get the following errors in the log:
Caused by: java.lang.SecurityException: Intercepted System.exit(1) at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
then after that:
Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]
any solution and explanation why this is happening?
Created 05-27-2019 11:43 PM
Created 05-29-2019 03:11 AM
Sorry, but where I can find the workflow.xml and the job.properties files?
And the following is the Sqoop import command I am trying to execute:
sqoop import \ --connect 'jdbc:sqlserver://myURL;database=myDB' \ --username user --password pass \ --table BigDataTest -m 1 --check-column lastmodified \ --merge-key id \ --incremental lastmodified \ --compression-codec=snappy \ --target-dir /user/hive/warehouse/dwh_db.db/bigdatatest \ --hive-table bigDataTest \ --map-column-java lastmodified=String \ --class-name BigDataTest \ --fields-terminated-by '\001' --fields-terminated-by '\n'
Created on 06-12-2019 01:39 AM - edited 06-12-2019 01:43 AM
Hello Eric .... any solution for this?!
I have version CHD 6.2
And I am trying to run it from Hue
Created 06-12-2019 10:29 PM
Created 06-16-2019 01:56 AM
That is a very long log....
This forum does not allow more than 50K char.s ....
The following is the last 50K char.s of the log generated:
dfs.namenode.checkpoint.dir : file://${hadoop.tmp.dir}/dfs/namesecondary
dfs.webhdfs.rest-csrf.browser-useragents-regex : ^Mozilla.*,^Opera.*
dfs.namenode.top.windows.minutes : 1,5,25
dfs.client.use.legacy.blockreader.local : false
mapreduce.job.maxtaskfailures.per.tracker : 3
mapreduce.shuffle.max.connections : 0
net.topology.node.switch.mapping.impl : org.apache.hadoop.net.ScriptBasedMapping
hadoop.kerberos.keytab.login.autorenewal.enabled : false
yarn.client.application-client-protocol.poll-interval-ms : 200
mapreduce.fileoutputcommitter.marksuccessfuljobs : true
yarn.nodemanager.localizer.address : ${yarn.nodemanager.hostname}:8040
dfs.namenode.list.cache.pools.num.responses : 100
nfs.server.port : 2049
dfs.namenode.https-address.NameServiceOne.namenode417 : masternode:9871
hadoop.proxyuser.HTTP.hosts : *
dfs.checksum.type : CRC32C
fs.s3a.readahead.range : 64K
dfs.client.read.short.circuit.replica.stale.threshold.ms : 1800000
dfs.ha.namenodes.NameServiceOne : namenode417,namenode434
ha.zookeeper.parent-znode : /hadoop-ha
yarn.sharedcache.admin.thread-count : 1
yarn.nodemanager.resource.cpu-vcores : -1
mapreduce.jobhistory.http.policy : HTTP_ONLY
fs.s3a.attempts.maximum : 20
dfs.datanode.lazywriter.interval.sec : 60
yarn.log-aggregation.retain-check-interval-seconds : -1
yarn.resourcemanager.node-ip-cache.expiry-interval-secs : -1
yarn.timeline-service.client.fd-clean-interval-secs : 60
fs.wasbs.impl : org.apache.hadoop.fs.azure.NativeAzureFileSystem$Secure
dfs.federation.router.reader.count : 1
hadoop.ssl.keystores.factory.class : org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory
hadoop.zk.num-retries : 1000
mapreduce.job.split.metainfo.maxsize : 10000000
hadoop.security.random.device.file.path : /dev/urandom
yarn.client.nodemanager-connect.max-wait-ms : 180000
yarn.app.mapreduce.client-am.ipc.max-retries : 3
dfs.namenode.snapshotdiff.allow.snap-root-descendant : true
yarn.nodemanager.container-diagnostics-maximum-size : 10000
yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage : false
dfs.namenode.ec.system.default.policy : RS-6-3-1024k
dfs.replication.max : 512
dfs.datanode.https.address : 0.0.0.0:9865
dfs.ha.standby.checkpoints : true
ipc.client.kill.max : 10
mapreduce.job.committer.setup.cleanup.needed : true
dfs.client.domain.socket.data.traffic : false
yarn.nodemanager.localizer.cache.target-size-mb : 10240
yarn.resourcemanager.admin.client.thread-count : 1
hadoop.security.group.mapping.ldap.connection.timeout.ms : 60000
yarn.timeline-service.store-class : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore
yarn.resourcemanager.nm-container-queuing.queue-limit-stdev : 1.0f
yarn.resourcemanager.zk-appid-node.split-index : 0
hadoop.tmp.dir : /tmp/hadoop-${user.name}
dfs.domain.socket.disable.interval.seconds : 1
fs.s3a.etag.checksum.enabled : false
hadoop.security.kms.client.failover.sleep.base.millis : 100
yarn.node-labels.configuration-type : centralized
fs.s3a.retry.interval : 500ms
dfs.datanode.http.internal-proxy.port : 0
yarn.timeline-service.ttl-ms : 604800000
mapreduce.task.exit.timeout.check-interval-ms : 20000
oozie.sqoop.args.7 : \
--table
oozie.sqoop.args.8 : category
mapreduce.map.speculative : false
oozie.sqoop.args.5 : --password
oozie.sqoop.args.6 : myUsername
yarn.nodemanager.linux-container-executor.cgroups.delete-timeout-ms : 1000
yarn.timeline-service.recovery.enabled : false
oozie.sqoop.args.9 : -m
yarn.nodemanager.recovery.dir : ${hadoop.tmp.dir}/yarn-nm-recovery
mapreduce.job.counters.max : 120
dfs.namenode.name.cache.threshold : 10
oozie.sqoop.args.0 : import
dfs.namenode.caching.enabled : true
dfs.namenode.max.full.block.report.leases : 6
oozie.sqoop.args.3 : \
--username
yarn.nodemanager.linux-container-executor.cgroups.delete-delay-ms : 20
dfs.namenode.max.extra.edits.segments.retained : 10000
oozie.sqoop.args.4 : myUsername
dfs.webhdfs.user.provider.user.pattern : ^[A-Za-z_][A-Za-z0-9._-]*[$]?$
yarn.webapp.ui2.enable : false
oozie.sqoop.args.1 : \
--connect
oozie.sqoop.args.2 : 'jdbc:sqlserver://myServer;database=myDB'
dfs.client.mmap.enabled : true
mapreduce.map.log.level : INFO
dfs.datanode.ec.reconstruction.threads : 8
hadoop.fuse.timer.period : 5
yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms : 1000
hadoop.zk.timeout-ms : 10000
ha.health-monitor.check-interval.ms : 1000
dfs.client.hedged.read.threshold.millis : 500
yarn.resourcemanager.fs.state-store.retry-interval-ms : 1000
mapreduce.output.fileoutputformat.compress : false
yarn.sharedcache.store.in-memory.staleness-period-mins : 10080
dfs.client.write.byte-array-manager.count-limit : 2048
mapreduce.application.framework.path : hdfs://NameServiceOne//user/yarn/mapreduce/mr-framework/3.0.0-cdh6.2.0-mr-framework.tar.gz#mr-framework
hadoop.security.group.mapping.providers.combined : true
fs.AbstractFileSystem.har.impl : org.apache.hadoop.fs.HarFs
mapreduce.job.running.map.limit : 0
yarn.nodemanager.webapp.address : ${yarn.nodemanager.hostname}:8042
mapreduce.reduce.input.buffer.percent : 0.0
mapreduce.job.cache.files : hdfs://NameServiceOne/user/hue/oozie/deployments/_admin_-oozie-312-1560674439.04/lib/hive-site.xml#hive-site.xml,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-exec-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-security-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slider-core-0.90.2-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-repackaged-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcodings-1.0.18.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fst-2.50.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ST4-4.0.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-protobuf.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-0.23.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ecj-4.4.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libthrift-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-crypto-1.0.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-schemas-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/transaction-api-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-webapp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joda-time-2.9.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-core-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-hcatalog-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-tez.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-commons-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-container-servlet-core-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-java7-2.4.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-utils-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-hive.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stringtemplate-3.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-core-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bonecp-0.8.0.RELEASE.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-core-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/htrace-core4-4.1.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-api-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-client-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-encoding.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-ipc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-cli.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fastutil-7.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guava-11.0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oro-2.0.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-api-2.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-tree-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcpkix-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-pool-1.5.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jpam-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/osgi-resource-locator-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-annotations-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpcore-4.4.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-archives.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slf4j-api-1.7.25.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-hbase-compat-1.0-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hsqldb-1.8.0.10.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-resourcemanager.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jaas-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-linq4j-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-codec-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/gson-2.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol-shaded.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-api-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jta-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-dbcp-1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-common-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aggdesigner-algorithm-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-json-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/opencsv-2.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-column.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.ws.rs-api-2.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-locator-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javolution-5.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-web-proxy.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-ant.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/groovy-all-2.4.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop-bundle.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-http-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-api-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/mssql-jdbc-6.2.1.jre7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-common-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joni-2.1.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-2.7.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jstl-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jndi-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libfb303-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-java-1.1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.el-3.0.1-b11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-plus-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-httpclient-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-netty.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/re2j-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-io-2.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javassist-3.20.0-GA.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-media-jaxb-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-curator-2.12.0.pom#apache-curator-2.12.0.pom,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-classification.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-databind-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-replication.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-storage-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet-api-3.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-client-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.annotation-api-1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-framework-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-scheduler.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcprov-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-procedure.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-rewrite-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-serde.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-spec-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-druid-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.jdo-3.2.0-m3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-core-4.1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-jvm-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/logredactor-2.0.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-web-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jsp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/java-util-1.9.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcommander-1.30.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-metastore.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-runner-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-2.6.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compress-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-assistedinject-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-slf4j-impl-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ehcache-3.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop2-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpclient-4.5.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/objenesis-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-common-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-server-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/velocity-1.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/janino-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-http.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stax-api-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang3-3.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avatica-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-runtime-3.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-registry.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/leveldbjni-all-1.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-util-ajax-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jdo-api-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jline-2.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-impl-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-asl-1.9.13.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-io-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-launcher-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/audience-annotations-0.5.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-hadoop-compatibility.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-2.3.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/derby-10.14.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/findbugs-annotations-1.3.9-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/xz-1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/geronimo-jcache_1.0_spec-1.0-alpha-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-api-jdo-4.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/netty-3.10.6.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-xml-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ivy-2.4.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/paranamer-2.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compiler-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-miscellaneous.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-mapred-hadoop2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-applicationhistoryservice.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-rdbms-4.1.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-core-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service-rpc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jsr305-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-mapper-asl-1.9.13-cloudera.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-20090211.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/validation-api-1.1.0.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-guava-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-zookeeper-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-format.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2.17.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/dropwizard-metrics-hadoop-metrics2-reporter-0.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-annotations-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-io-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-jackson.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/disruptor-3.3.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-orc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie.jar
dfs.client.slow.io.warning.threshold.ms : 30000
fs.s3a.multipart.size : 100M
yarn.app.mapreduce.am.job.committer.commit-window : 10000
dfs.qjournal.new-epoch.timeout.ms : 120000
yarn.timeline-service.webapp.rest-csrf.enabled : false
hadoop.proxyuser.flume.hosts : *
dfs.namenode.edits.asynclogging : true
yarn.timeline-service.reader.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineReaderImpl
yarn.app.mapreduce.am.staging-dir.erasurecoding.enabled : false
dfs.blockreport.incremental.intervalMsec : 0
dfs.datanode.network.counts.cache.max.size : 2147483647
dfs.namenode.https-address.NameServiceOne.namenode434 : node3:9871
yarn.timeline-service.writer.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineWriterImpl
mapreduce.ifile.readahead : true
dfs.qjournal.get-journal-state.timeout.ms : 120000
yarn.timeline-service.entity-group-fs-store.summary-store : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore
dfs.client.socketcache.capacity : 16
fs.s3a.s3guard.ddb.table.create : false
dfs.client.retry.policy.spec : 10000,6,60000,10
mapreduce.output.fileoutputformat.compress.codec : org.apache.hadoop.io.compress.DefaultCodec
fs.s3a.socket.recv.buffer : 8192
dfs.datanode.fsdatasetcache.max.threads.per.volume : 4
dfs.namenode.reencrypt.batch.size : 1000
yarn.sharedcache.store.in-memory.initial-delay-mins : 10
mapreduce.jobhistory.webapp.address : masternode:19888
fs.adl.impl : org.apache.hadoop.fs.adl.AdlFileSystem
fs.AbstractFileSystem.gs.impl : com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS
mapreduce.task.userlog.limit.kb : 0
fs.s3a.connection.ssl.enabled : true
yarn.router.rmadmin.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.rmadmin.DefaultRMAdminRequestInterceptor
yarn.sharedcache.webapp.address : 0.0.0.0:8788
hadoop.fuse.connection.timeout : 300
dfs.http.client.retry.policy.spec : 10000,6,60000,10
yarn.resourcemanager.rm.container-allocation.expiry-interval-ms : 600000
ipc.server.max.connections : 0
yarn.app.mapreduce.am.resource.mb : 3072
hadoop.security.groups.cache.secs : 300
dfs.datanode.peer.stats.enabled : false
dfs.replication : 3
mapreduce.shuffle.transfer.buffer.size : 131072
dfs.namenode.audit.log.async : false
hadoop.security.group.mapping.ldap.directory.search.timeout : 10000
dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold : 10737418240
dfs.datanode.disk.check.timeout : 10m
yarn.app.mapreduce.client-am.ipc.max-retries-on-timeouts : 3
fs.s3a.committer.threads : 8
dfs.checksum.combine.mode : MD5MD5CRC
yarn.resourcemanager.nodemanager-graceful-decommission-timeout-secs : 3600
yarn.scheduler.maximum-allocation-vcores : 6
yarn.nodemanager.sleep-delay-before-sigkill.ms : 250
fs.AbstractFileSystem.abfs.impl : org.apache.hadoop.fs.azurebfs.Abfs
mapreduce.job.acl-modify-job :
fs.automatic.close : true
fs.azure.sas.expiry.period : 90d
dfs.qjm.operations.timeout : 60s
hadoop.proxyuser.httpfs.hosts : *
dfs.namenode.stale.datanode.minimum.interval : 3
dfs.namenode.reencrypt.edek.threads : 10
dfs.federation.router.store.membership.expiration : 300000
hadoop.security.groups.cache.background.reload.threads : 3
mapreduce.input.fileinputformat.list-status.num-threads : 1
hadoop.security.group.mapping.ldap.posix.attr.gid.name : gidNumber
dfs.namenode.acls.enabled : false
dfs.client.short.circuit.replica.stale.threshold.ms : 1800000
dfs.namenode.resource.du.reserved : 104857600
dfs.federation.router.connection.clean.ms : 10000
dfs.client.server-defaults.validity.period.ms : 3600000
dfs.federation.router.metrics.class : org.apache.hadoop.hdfs.server.federation.metrics.FederationRPCPerformanceMonitor
mapreduce.shuffle.listen.queue.size : 128
mapreduce.jobhistory.intermediate-done-dir : ${yarn.app.mapreduce.am.staging-dir}/history/done_intermediate
mapreduce.client.libjars.wildcard : true
dfs.federation.router.cache.ttl : 60000
yarn.nodemanager.recovery.compaction-interval-secs : 3600
dfs.namenode.edits.noeditlogchannelflush : false
mapreduce.reduce.shuffle.input.buffer.percent : 0.70
yarn.http.policy : HTTP_ONLY
mapreduce.map.maxattempts : 4
dfs.namenode.audit.loggers : default
io.serializations : org.apache.hadoop.io.serializer.WritableSerialization, org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization, org.apache.hadoop.io.serializer.avro.AvroReflectSerialization
hadoop.security.groups.cache.warn.after.ms : 5000
dfs.client.write.byte-array-manager.count-reset-time-period-ms : 10000
yarn.nodemanager.webapp.rest-csrf.custom-header : X-XSRF-Header
yarn.app.mapreduce.am.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH
dfs.namenode.snapshot.capture.openfiles : true
yarn.node-labels.fs-store.impl.class : org.apache.hadoop.yarn.nodelabels.FileSystemNodeLabelsStore
hadoop.http.cross-origin.allowed-methods : GET,POST,HEAD
dfs.qjournal.queued-edits.limit.mb : 10
mapreduce.jobhistory.webapp.rest-csrf.enabled : false
dfs.http.policy : HTTP_ONLY
dfs.balancer.max-size-to-move : 10737418240
dfs.datanode.sync.behind.writes.in.background : false
hadoop.zk.acl : world:anyone:rwcda
yarn.nodemanager.container.stderr.pattern : {*stderr*,*STDERR*}
dfs.namenode.reencrypt.throttle.limit.updater.ratio : 1.0
mapreduce.cluster.local.dir : ${hadoop.tmp.dir}/mapred/local
hadoop.kerberos.kinit.command : kinit
dfs.namenode.secondary.https-address : 0.0.0.0:9869
dfs.namenode.metrics.logger.period.seconds : 600
dfs.block.access.token.lifetime : 600
dfs.ha.automatic-failover.enabled.NameServiceOne : true
dfs.namenode.delegation.token.max-lifetime : 604800000
dfs.datanode.drop.cache.behind.writes : false
dfs.mover.address : 0.0.0.0:0
dfs.block.placement.ec.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyRackFaultTolerant
dfs.namenode.num.extra.edits.retained : 1000000
ipc.client.connect.max.retries.on.timeouts : 45
fs.viewfs.rename.strategy : SAME_MOUNTPOINT
fs.client.resolve.topology.enabled : false
hadoop.proxyuser.hive.hosts : *
yarn.resourcemanager.node-labels.provider.fetch-interval-ms : 1800000
yarn.nodemanager.container-metrics.enable : true
mapreduce.job.map.output.collector.class : org.apache.hadoop.mapred.MapTask$MapOutputBuffer
fs.s3a.fast.upload.buffer : disk
ha.health-monitor.connect-retry-interval.ms : 1000
dfs.namenode.edekcacheloader.initial.delay.ms : 3000
dfs.edit.log.transfer.bandwidthPerSec : 0
dfs.ha.tail-edits.in-progress : false
dfs.federation.router.heartbeat.interval : 5000
ssl.client.truststore.reload.interval : 10000
dfs.client.datanode-restart.timeout : 30s
io.mapfile.bloom.size : 1048576
hadoop.security.kms.client.authentication.retry-count : 1
dfs.client-write-packet-size : 65536
fs.ftp.data.connection.mode : ACTIVE_LOCAL_DATA_CONNECTION_MODE
fs.swift.impl : org.apache.hadoop.fs.swift.snative.SwiftNativeFileSystem
yarn.resourcemanager.webapp.rest-csrf.methods-to-ignore : GET,OPTIONS,HEAD
mapreduce.job.max.map : -1
yarn.app.mapreduce.shuffle.log.backups : 0
ftp.blocksize : 67108864
dfs.namenode.kerberos.principal.pattern : *
yarn.resourcemanager.scheduler.monitor.enable : false
dfs.webhdfs.socket.connect-timeout : 60s
dfs.namenode.replication.max-streams : 2
nfs.allow.insecure.ports : true
yarn.sharedcache.nm.uploader.thread-count : 20
dfs.federation.router.admin.enable : true
yarn.app.mapreduce.client.job.retry-interval : 2000
yarn.scheduler.configuration.store.max-logs : 1000
hadoop.security.authorization : false
yarn.timeline-service.version : 1.0f
yarn.am.liveness-monitor.expiry-interval-ms : 600000
fs.har.impl.disable.cache : true
hadoop.proxyuser.hdfs.hosts : *
mapreduce.job.reduce.slowstart.completedmaps : 0.8
yarn.timeline-service.leveldb-timeline-store.path : ${hadoop.tmp.dir}/yarn/timeline
dfs.namenode.upgrade.domain.factor : ${dfs.replication}
mapreduce.jobhistory.minicluster.fixed.ports : false
mapreduce.application.classpath : $HADOOP_CLIENT_CONF_DIR,$PWD/mr-framework/*,$MR2_CLASSPATH
yarn.resourcemanager.delegation.token.max-lifetime : 604800000
yarn.resourcemanager.ha.automatic-failover.enabled : true
mapreduce.reduce.java.opts : -Djava.net.preferIPv4Stack=true
dfs.datanode.socket.write.timeout : 480000
dfs.namenode.accesstime.precision : 3600000
dfs.namenode.redundancy.considerLoad.factor : 2.0
yarn.resourcemanager.store.class : org.apache.hadoop.yarn.server.resourcemanager.recovery.FileSystemRMStateStore
io.mapfile.bloom.error.rate : 0.005
yarn.nodemanager.webapp.rest-csrf.enabled : false
yarn.timeline-service.leveldb-state-store.path : ${hadoop.tmp.dir}/yarn/timeline
hadoop.proxyuser.hive.groups : *
dfs.federation.router.rpc-address : 0.0.0.0:8888
fs.s3a.committer.staging.unique-filenames : true
dfs.namenode.support.allow.format : true
yarn.scheduler.configuration.zk-store.parent-path : /confstore
dfs.content-summary.limit : 5000
yarn.timeline-service.writer.flush-interval-seconds : 60
yarn.nodemanager.container-executor.class : org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor
dfs.namenode.posix.acl.inheritance.enabled : true
dfs.datanode.outliers.report.interval : 30m
hadoop.security.kms.client.encrypted.key.cache.low-watermark : 0.3f
dfs.namenode.top.enabled : true
yarn.app.mapreduce.shuffle.log.separate : true
hadoop.user.group.static.mapping.overrides : dr.who=;
dfs.federation.router.http-address : 0.0.0.0:50071
fs.s3a.retry.throttle.interval : 1000ms
yarn.nodemanager.amrmproxy.address : 0.0.0.0:8049
mapreduce.jobhistory.webapp.rest-csrf.custom-header : X-XSRF-Header
yarn.webapp.xfs-filter.enabled : true
dfs.client.cached.conn.retry : 3
dfs.client.key.provider.cache.expiry : 864000000
dfs.namenode.path.based.cache.refresh.interval.ms : 30000
yarn.nodemanager.collector-service.thread-count : 5
dfs.block.replicator.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyDefault
dfs.namenode.fs-limits.max-directory-items : 1048576
dfs.ha.log-roll.period : 120s
yarn.nodemanager.runtime.linux.docker.capabilities : CHOWN,DAC_OVERRIDE,FSETID,FOWNER,MKNOD,NET_RAW,SETGID,SETUID,SETFCAP,SETPCAP,NET_BIND_SERVICE,SYS_CHROOT,KILL,AUDIT_WRITE
yarn.nodemanager.distributed-scheduling.enabled : false
ipc.client.fallback-to-simple-auth-allowed : false
yarn.minicluster.fixed.ports : false
yarn.nodemanager.remote-app-log-dir : /tmp/logs
yarn.timeline-service.entity-group-fs-store.scan-interval-seconds : 60
dfs.xframe.enabled : true
yarn.nodemanager.resource.percentage-physical-cpu-limit : 100
mapreduce.job.tags : oozie-564a124254f1fd53cb03553181f7e603
dfs.namenode.fs-limits.max-xattr-size : 16384
dfs.datanode.http.address : 0.0.0.0:9864
dfs.namenode.blocks.per.postponedblocks.rescan : 10000
fs.s3a.s3guard.cli.prune.age : 86400000
dfs.web.authentication.filter : org.apache.hadoop.hdfs.web.AuthFilter
dfs.namenode.maintenance.replication.min : 1
hadoop.jetty.logs.serve.aliases : true
dfs.webhdfs.ugi.expire.after.access : 600000
dfs.namenode.max.op.size : 52428800
mapreduce.jobhistory.admin.acl : *
mapreduce.job.reducer.unconditional-preempt.delay.sec : 300
yarn.app.mapreduce.am.hard-kill-timeout-ms : 10000
yarn.resourcemanager.display.per-user-apps : false
yarn.resourcemanager.node-removal-untracked.timeout-ms : 60000
yarn.resourcemanager.webapp.address : masternode:8088
mapreduce.jobhistory.recovery.enable : false
yarn.sharedcache.store.in-memory.check-period-mins : 720
dfs.client.test.drop.namenode.response.number : 0
fs.df.interval : 60000
fs.s3a.assumed.role.session.duration : 30m
mapreduce.job.cache.limit.max-single-resource-mb : 0
yarn.timeline-service.enabled : false
dfs.disk.balancer.block.tolerance.percent : 10
dfs.webhdfs.netty.high.watermark : 65535
mapreduce.task.profile : false
hadoop.http.cross-origin.allowed-headers : X-Requested-With,Content-Type,Accept,Origin
yarn.router.webapp.address : 0.0.0.0:8089
dfs.datanode.balance.max.concurrent.moves : 50
yarn.nodemanager.hostname : 0.0.0.0
mapreduce.task.exit.timeout : 60000
yarn.resourcemanager.nm-container-queuing.max-queue-length : 15
mapreduce.job.token.tracking.ids.enabled : false
yarn.scheduler.increment-allocation-mb : 512
fs.s3a.assumed.role.credentials.provider : org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider
fs.azure.authorization.caching.enable : true
hadoop.security.kms.client.failover.sleep.max.millis : 2000
dfs.client.mmap.retry.timeout.ms : 300000
yarn.resourcemanager.webapp.rest-csrf.custom-header : X-XSRF-Header
yarn.resourcemanager.nm-container-queuing.max-queue-wait-time-ms : 100
mapreduce.jobhistory.move.thread-count : 3
dfs.permissions.enabled : true
fs.AbstractFileSystem.hdfs.impl : org.apache.hadoop.fs.Hdfs
yarn.nodemanager.container-localizer.log.level : INFO
hadoop.http.filter.initializers : org.apache.hadoop.http.lib.StaticUserWebFilter
yarn.timeline-service.http-authentication.simple.anonymous.allowed : true
yarn.nodemanager.runtime.linux.docker.allowed-container-networks : host,none,bridge
dfs.qjournal.accept-recovery.timeout.ms : 120000
yarn.sharedcache.client-server.thread-count : 50
fs.s3a.s3guard.ddb.max.retries : 9
fs.s3a.committer.magic.enabled : false
yarn.resourcemanager.resource-tracker.address : masternode:8031
mapreduce.jobhistory.jobname.limit : 50
dfs.domain.socket.path : /var/run/hdfs-sockets/dn
dfs.namenode.decommission.blocks.per.interval : 500000
dfs.qjournal.write-txns.timeout.ms : 20000
rpc.metrics.quantile.enable : false
yarn.federation.subcluster-resolver.class : org.apache.hadoop.yarn.server.federation.resolver.DefaultSubClusterResolverImpl
dfs.namenode.read-lock-reporting-threshold-ms : 5000
mapreduce.task.timeout : 600000
yarn.nodemanager.resource.memory-mb : -1
dfs.datanode.failed.volumes.tolerated : 0
yarn.nodemanager.disk-health-checker.min-healthy-disks : 0.25
mapreduce.framework.name : yarn
mapreduce.fileoutputcommitter.algorithm.version : 2
yarn.router.clientrm.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.clientrm.DefaultClientRequestInterceptor
yarn.sharedcache.nested-level : 3
fs.s3a.connection.timeout : 200000
hadoop.caller.context.signature.max.size : 40
hadoop.security.dns.log-slow-lookups.enabled : false
mapreduce.jobhistory.webapp.https.address : masternode:19890
file.client-write-packet-size : 65536
fs.s3a.s3guard.ddb.table.capacity.read : 500
ipc.client.ping : true
hadoop.proxyuser.oozie.hosts : *
yarn.resourcemanager.delayed.delegation-token.removal-interval-ms : 30000
dfs.client.failover.max.attempts : 15
dfs.balancer.max-no-move-interval : 60000
yarn.nodemanager.opportunistic-containers-use-pause-for-preemption : false
yarn.nodemanager.webapp.cross-origin.enabled : false
mapreduce.job.encrypted-intermediate-data : false
dfs.client.read.shortcircuit.streams.cache.expiry.ms : 300000
yarn.minicluster.control-resource-monitoring : false
dfs.disk.balancer.enabled : false
dfs.webhdfs.oauth2.enabled : false
yarn.nodemanager.health-checker.script.timeout-ms : 1200000
yarn.resourcemanager.fs.state-store.num-retries : 0
hadoop.ssl.require.client.cert : false
mapreduce.jobhistory.keytab : /etc/security/keytab/jhs.service.keytab
hadoop.security.uid.cache.secs : 14400
yarn.resourcemanager.ha.automatic-failover.zk-base-path : /yarn-leader-election
yarn.intermediate-data-encryption.enable : false
mapreduce.job.speculative.speculative-cap-running-tasks : 0.1
dfs.datanode.block.id.layout.upgrade.threads : 12
dfs.client.context : default
yarn.system-metrics-publisher.enabled : false
dfs.namenode.delegation.token.renew-interval : 86400000
yarn.timeline-service.entity-group-fs-store.app-cache-size : 10
fs.AbstractFileSystem.s3a.impl : org.apache.hadoop.fs.s3a.S3A
mapreduce.job.redacted-properties : fs.s3a.access.key,fs.s3a.secret.key,fs.adl.oauth2.credential,dfs.adls.oauth2.credential,fs.azure.account.oauth2.client.secret
yarn.client.load.resource-types.from-server : false
ipc.client.tcpnodelay : true
hadoop.proxyuser.httpfs.groups : *
yarn.resourcemanager.metrics.runtime.buckets : 60,300,1440
dfs.blockreport.intervalMsec : 21600000
dfs.datanode.oob.timeout-ms : 1500,0,0,0
yarn.client.application-client-protocol.poll-timeout-ms : -1
zlib.compress.level : DEFAULT_COMPRESSION
mapreduce.job.sharedcache.mode : disabled
io.map.index.skip : 0
mapreduce.job.hdfs-servers : ${fs.defaultFS}
mapreduce.map.output.compress : true
hadoop.security.kms.client.encrypted.key.cache.num.refill.threads : 2
dfs.namenode.edekcacheloader.interval.ms : 1000
mapreduce.task.merge.progress.records : 10000
yarn.nodemanager.aux-services.mapreduce_shuffle.class : org.apache.hadoop.mapred.ShuffleHandler
dfs.namenode.missing.checkpoint.periods.before.shutdown : 3
tfile.fs.output.buffer.size : 262144
dfs.client.failover.connection.retries : 0
fs.du.interval : 600000
dfs.edit.log.transfer.timeout : 30000
dfs.namenode.top.window.num.buckets : 10
dfs.data.transfer.server.tcpnodelay : true
hadoop.zk.retry-interval-ms : 1000
yarn.sharedcache.uploader.server.address : 0.0.0.0:8046
dfs.http.client.failover.max.attempts : 15
fs.s3a.socket.send.buffer : 8192
dfs.client.block.write.locateFollowingBlock.retries : 7
hadoop.registry.zk.quorum : localhost:2181
mapreduce.jvm.system-properties-to-log : os.name,os.version,java.home,java.runtime.version,java.vendor,java.version,java.vm.name,java.class.path,java.io.tmpdir,user.dir,user.name
hadoop.http.cross-origin.allowed-origins : *
dfs.namenode.enable.retrycache : true
dfs.datanode.du.reserved : 0
hadoop.registry.system.acls : sasl:yarn@, sasl:mapred@, sasl:hdfs@
yarn.nodemanager.resource-plugins.gpu.docker-plugin.nvidia-docker-v1.endpoint : http://localhost:3476/v1.0/docker/cli
mapreduce.job.encrypted-intermediate-data.buffer.kb : 128
dfs.data.transfer.client.tcpnodelay : true
yarn.resourcemanager.webapp.xfs-filter.xframe-options : SAMEORIGIN
mapreduce.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH
mapreduce.task.profile.reduce.params : ${mapreduce.task.profile.params}
mapreduce.reduce.memory.mb : 0
hadoop.caller.context.enabled : false
hadoop.http.authentication.kerberos.principal : HTTP/_HOST@LOCALHOST
yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb : 0
dfs.qjournal.prepare-recovery.timeout.ms : 120000
dfs.datanode.transferTo.allowed : true
oozie.action.rootlogger.log.level : INFO
hadoop.security.sensitive-config-keys :
secret$
password$
ssl.keystore.pass$
fs.s3.*[Ss]ecret.?[Kk]ey
fs.s3a.*.server-side-encryption.key
fs.azure.account.key.*
credential$
oauth.*secret
oauth.*password
oauth.*token
hadoop.security.sensitive-config-keys
mapreduce.client.completion.pollinterval : 5000
dfs.namenode.name.dir.restore : false
dfs.namenode.full.block.report.lease.length.ms : 300000
dfs.namenode.secondary.http-address : 0.0.0.0:9868
hadoop.http.logs.enabled : true
hadoop.security.group.mapping.ldap.read.timeout.ms : 60000
yarn.resourcemanager.max-log-aggregation-diagnostics-in-memory : 10
dfs.namenode.delegation.token.always-use : false
yarn.resourcemanager.webapp.https.address : masternode:8090
fs.s3a.retry.throttle.limit : ${fs.s3a.attempts.maximum}
dfs.client.read.striped.threadpool.size : 18
mapreduce.job.cache.limit.max-resources : 0
hadoop.proxyuser.HTTP.groups : *
--------------------
Setting up log4j2
log4j2 configuration file created at /yarn/nm/usercache/admin/appcache/application_1560674082717_0001/container_1560674082717_0001_01_000001/sqoop-log4j2.xml
Sqoop command arguments :
import
\
--connect
'jdbc:sqlserver://myServer;database=myDB'
\
--username
myUsername
--password
********
\
--table
category
-m
1
--check-column
LastEditOn
\
--merge-key
'Reference
ID'
\
--incremental
lastmodified
\
--compression-codec=snappy
\
--target-dir
/user/hive/warehouse/myDB.db/category
\
--hive-table
category
\
--map-column-hive
LastEditOn=timestamp,CreatedOn=timestamp
\
--fields-terminated-by
'\001'
--fields-terminated-by
'\n'
Fetching child yarn jobs
tag id : oozie-564a124254f1fd53cb03553181f7e603
No child applications found
=================================================================
>>> Invoking Sqoop command line now >>>
<<< Invocation of Sqoop command completed <<<
No child hadoop job is executed.
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.oozie.action.hadoop.LauncherAM.runActionMain(LauncherAM.java:410)
at org.apache.oozie.action.hadoop.LauncherAM.access$300(LauncherAM.java:55)
at org.apache.oozie.action.hadoop.LauncherAM$2.run(LauncherAM.java:223)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
at org.apache.oozie.action.hadoop.LauncherAM.run(LauncherAM.java:217)
at org.apache.oozie.action.hadoop.LauncherAM$1.run(LauncherAM.java:153)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
at org.apache.oozie.action.hadoop.LauncherAM.main(LauncherAM.java:141)
Caused by: java.lang.SecurityException: Intercepted System.exit(1)
at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
at java.lang.Runtime.exit(Runtime.java:107)
at java.lang.System.exit(System.java:971)
at org.apache.sqoop.Sqoop.main(Sqoop.java:252)
at org.apache.oozie.action.hadoop.SqoopMain.runSqoopJob(SqoopMain.java:214)
at org.apache.oozie.action.hadoop.SqoopMain.run(SqoopMain.java:199)
at org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:104)
at org.apache.oozie.action.hadoop.SqoopMain.main(SqoopMain.java:51)
... 16 more
Intercepting System.exit(1)
Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]
Oozie Launcher, uploading action data to HDFS sequence file: hdfs://NameServiceOne/user/admin/oozie-oozi/0000000-190616123600049-oozie-oozi-W/sqoop-c9e7--sqoop/action-data.seq
12:41:09.783 [main] INFO org.apache.hadoop.io.compress.CodecPool - Got brand-new compressor [.deflate]
Stopping AM
12:41:09.983 [main] INFO org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl - Waiting for application to be successfully unregistered.
Callback notification attempts left 0
Callback notification trying http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED
Callback notification to http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED succeeded
Callback notification succeeded
Created 06-16-2019 05:35 PM
Created 06-19-2019 01:03 AM
I am sorry but there is nothing wrong with the syntax, as if I run it on the terminal it completes successfully.
I have a doubt regarding security, and that is because of the following line in the log
aused by: java.lang.SecurityException: Intercepted System.exit(1) at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
I would like to note also that this is the first try to run sqoop script on Hue after a fresh installation of cdh 6.2... so I am afraid there is something that I've missed in the configuration, but I really can't find it 😞
Created 06-19-2019 04:04 PM
Created 06-19-2019 11:31 PM
I hope this is what you want for the workflow:
<workflow-app name="Batch job for query-sqoop1" xmlns="uri:oozie:workflow:0.5">
<start to="sqoop-fde5"/>
<kill name="Kill">
<message>Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<action name="sqoop-fde5">
<sqoop xmlns="uri:oozie:sqoop-action:0.2">
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node>
<command>import \
--connect 'jdbc:sqlserver://11.11.11.11;database=DBXYZ' \
--username theUser --password thePassword \
--table category -m 1 --check-column LastEditOn \
--merge-key 'Reference ID' \
--incremental lastmodified \
--compression-codec=snappy \
--target-dir /user/hive/warehouse/dwh_db_atlas_jrtf.db/category \
--hive-table category \
--map-column-hive LastEditOn=timestamp,CreatedOn=timestamp \
--fields-terminated-by '\001' --fields-terminated-by '\n'</command>
</sqoop>
<ok to="End"/>
<error to="Kill"/>
</action>
<end name="End"/>
</workflow-app>For the job configuration, I am really not sure where to find it. The one I can reach is a lot of scrolls that can't be taken screenshot in anyway.... so would you please give me the path to the job configuration ?