- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Sqoop job is being KILLED when executed though hue
- Labels:
-
Apache Sqoop
-
Cloudera Hue
Created on 05-27-2019 04:03 AM - edited 09-16-2022 07:24 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hello,
When I try to run a Sqoop import command through Hue, the job is KILLED always, and I get the following errors in the log:
Caused by: java.lang.SecurityException: Intercepted System.exit(1) at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
then after that:
Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]
any solution and explanation why this is happening?
Created 05-27-2019 11:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Can you please share the content of workflow.xml and job.properties files here for me to have a look? Do you use "--query" options in Sqoop?
Also, what version of CDH are you using?
Cheers
Eric
Created 05-29-2019 03:11 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Sorry, but where I can find the workflow.xml and the job.properties files?
And the following is the Sqoop import command I am trying to execute:
sqoop import \ --connect 'jdbc:sqlserver://myURL;database=myDB' \ --username user --password pass \ --table BigDataTest -m 1 --check-column lastmodified \ --merge-key id \ --incremental lastmodified \ --compression-codec=snappy \ --target-dir /user/hive/warehouse/dwh_db.db/bigdatatest \ --hive-table bigDataTest \ --map-column-java lastmodified=String \ --class-name BigDataTest \ --fields-terminated-by '\001' --fields-terminated-by '\n'
Created on 06-12-2019 01:39 AM - edited 06-12-2019 01:43 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hello Eric .... any solution for this?!
I have version CHD 6.2
And I am trying to run it from Hue
Created 06-12-2019 10:29 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Since you are not using "--query" option, can you share the full error from the launcher log for review?
Just trying to get a bigger picture on the issue by checking what happened before and after.
Cheers
Eric
Created 06-16-2019 01:56 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
That is a very long log....
This forum does not allow more than 50K char.s ....
The following is the last 50K char.s of the log generated:
dfs.namenode.checkpoint.dir : file://${hadoop.tmp.dir}/dfs/namesecondary dfs.webhdfs.rest-csrf.browser-useragents-regex : ^Mozilla.*,^Opera.* dfs.namenode.top.windows.minutes : 1,5,25 dfs.client.use.legacy.blockreader.local : false mapreduce.job.maxtaskfailures.per.tracker : 3 mapreduce.shuffle.max.connections : 0 net.topology.node.switch.mapping.impl : org.apache.hadoop.net.ScriptBasedMapping hadoop.kerberos.keytab.login.autorenewal.enabled : false yarn.client.application-client-protocol.poll-interval-ms : 200 mapreduce.fileoutputcommitter.marksuccessfuljobs : true yarn.nodemanager.localizer.address : ${yarn.nodemanager.hostname}:8040 dfs.namenode.list.cache.pools.num.responses : 100 nfs.server.port : 2049 dfs.namenode.https-address.NameServiceOne.namenode417 : masternode:9871 hadoop.proxyuser.HTTP.hosts : * dfs.checksum.type : CRC32C fs.s3a.readahead.range : 64K dfs.client.read.short.circuit.replica.stale.threshold.ms : 1800000 dfs.ha.namenodes.NameServiceOne : namenode417,namenode434 ha.zookeeper.parent-znode : /hadoop-ha yarn.sharedcache.admin.thread-count : 1 yarn.nodemanager.resource.cpu-vcores : -1 mapreduce.jobhistory.http.policy : HTTP_ONLY fs.s3a.attempts.maximum : 20 dfs.datanode.lazywriter.interval.sec : 60 yarn.log-aggregation.retain-check-interval-seconds : -1 yarn.resourcemanager.node-ip-cache.expiry-interval-secs : -1 yarn.timeline-service.client.fd-clean-interval-secs : 60 fs.wasbs.impl : org.apache.hadoop.fs.azure.NativeAzureFileSystem$Secure dfs.federation.router.reader.count : 1 hadoop.ssl.keystores.factory.class : org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory hadoop.zk.num-retries : 1000 mapreduce.job.split.metainfo.maxsize : 10000000 hadoop.security.random.device.file.path : /dev/urandom yarn.client.nodemanager-connect.max-wait-ms : 180000 yarn.app.mapreduce.client-am.ipc.max-retries : 3 dfs.namenode.snapshotdiff.allow.snap-root-descendant : true yarn.nodemanager.container-diagnostics-maximum-size : 10000 yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage : false dfs.namenode.ec.system.default.policy : RS-6-3-1024k dfs.replication.max : 512 dfs.datanode.https.address : 0.0.0.0:9865 dfs.ha.standby.checkpoints : true ipc.client.kill.max : 10 mapreduce.job.committer.setup.cleanup.needed : true dfs.client.domain.socket.data.traffic : false yarn.nodemanager.localizer.cache.target-size-mb : 10240 yarn.resourcemanager.admin.client.thread-count : 1 hadoop.security.group.mapping.ldap.connection.timeout.ms : 60000 yarn.timeline-service.store-class : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore yarn.resourcemanager.nm-container-queuing.queue-limit-stdev : 1.0f yarn.resourcemanager.zk-appid-node.split-index : 0 hadoop.tmp.dir : /tmp/hadoop-${user.name} dfs.domain.socket.disable.interval.seconds : 1 fs.s3a.etag.checksum.enabled : false hadoop.security.kms.client.failover.sleep.base.millis : 100 yarn.node-labels.configuration-type : centralized fs.s3a.retry.interval : 500ms dfs.datanode.http.internal-proxy.port : 0 yarn.timeline-service.ttl-ms : 604800000 mapreduce.task.exit.timeout.check-interval-ms : 20000 oozie.sqoop.args.7 : \ --table oozie.sqoop.args.8 : category mapreduce.map.speculative : false oozie.sqoop.args.5 : --password oozie.sqoop.args.6 : myUsername yarn.nodemanager.linux-container-executor.cgroups.delete-timeout-ms : 1000 yarn.timeline-service.recovery.enabled : false oozie.sqoop.args.9 : -m yarn.nodemanager.recovery.dir : ${hadoop.tmp.dir}/yarn-nm-recovery mapreduce.job.counters.max : 120 dfs.namenode.name.cache.threshold : 10 oozie.sqoop.args.0 : import dfs.namenode.caching.enabled : true dfs.namenode.max.full.block.report.leases : 6 oozie.sqoop.args.3 : \ --username yarn.nodemanager.linux-container-executor.cgroups.delete-delay-ms : 20 dfs.namenode.max.extra.edits.segments.retained : 10000 oozie.sqoop.args.4 : myUsername dfs.webhdfs.user.provider.user.pattern : ^[A-Za-z_][A-Za-z0-9._-]*[$]?$ yarn.webapp.ui2.enable : false oozie.sqoop.args.1 : \ --connect oozie.sqoop.args.2 : 'jdbc:sqlserver://myServer;database=myDB' dfs.client.mmap.enabled : true mapreduce.map.log.level : INFO dfs.datanode.ec.reconstruction.threads : 8 hadoop.fuse.timer.period : 5 yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms : 1000 hadoop.zk.timeout-ms : 10000 ha.health-monitor.check-interval.ms : 1000 dfs.client.hedged.read.threshold.millis : 500 yarn.resourcemanager.fs.state-store.retry-interval-ms : 1000 mapreduce.output.fileoutputformat.compress : false yarn.sharedcache.store.in-memory.staleness-period-mins : 10080 dfs.client.write.byte-array-manager.count-limit : 2048 mapreduce.application.framework.path : hdfs://NameServiceOne//user/yarn/mapreduce/mr-framework/3.0.0-cdh6.2.0-mr-framework.tar.gz#mr-framework hadoop.security.group.mapping.providers.combined : true fs.AbstractFileSystem.har.impl : org.apache.hadoop.fs.HarFs mapreduce.job.running.map.limit : 0 yarn.nodemanager.webapp.address : ${yarn.nodemanager.hostname}:8042 mapreduce.reduce.input.buffer.percent : 0.0 mapreduce.job.cache.files : hdfs://NameServiceOne/user/hue/oozie/deployments/_admin_-oozie-312-1560674439.04/lib/hive-site.xml#hive-site.xml,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-exec-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-security-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slider-core-0.90.2-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-repackaged-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcodings-1.0.18.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fst-2.50.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ST4-4.0.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-protobuf.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-0.23.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ecj-4.4.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libthrift-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-crypto-1.0.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-schemas-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/transaction-api-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-webapp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joda-time-2.9.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-core-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-hcatalog-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-tez.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-commons-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-container-servlet-core-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-java7-2.4.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-core.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-utils-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-hive.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stringtemplate-3.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-core-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bonecp-0.8.0.RELEASE.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-core-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/htrace-core4-4.1.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-api-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-client-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-encoding.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-ipc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-cli.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/fastutil-7.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guava-11.0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oro-2.0.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-api-2.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/asm-tree-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcpkix-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-0.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-pool-1.5.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jpam-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/osgi-resource-locator-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-annotations-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpcore-4.4.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-archives.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/slf4j-api-1.7.25.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-hbase-compat-1.0-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hsqldb-1.8.0.10.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-data-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-resourcemanager.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jaas-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-linq4j-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-codec-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/gson-2.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol-shaded.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-api-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jta-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-dbcp-1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-common-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aggdesigner-algorithm-6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-json-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/opencsv-2.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-column.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.ws.rs-api-2.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hk2-locator-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javolution-5.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-web-proxy.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-ant.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/groovy-all-2.4.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-hadoop-bundle.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-protocol.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/aopalliance-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-http-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-api-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/mssql-jdbc-6.2.1.jre7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-common-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/joni-2.1.11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-2.7.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jstl-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-zookeeper.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-jndi-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/libfb303-0.9.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/snappy-java-1.1.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.el-3.0.1-b11.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-plus-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-httpclient-3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-netty.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/re2j-1.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-io-2.5.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javassist-3.20.0-GA.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-media-jaxb-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-curator-2.12.0.pom#apache-curator-2.12.0.pom,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-classification.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-databind-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-replication.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-storage-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet-api-3.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-client-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.annotation-api-1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/curator-framework-2.7.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-shims-scheduler.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-client.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/bcprov-jdk15on-1.60.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-procedure.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-rewrite-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-serde.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-spec-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/calcite-druid-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.jdo-3.2.0-m3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-core-4.1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/metrics-jvm-3.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/logredactor-2.0.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-web-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/apache-jsp-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/java-util-1.9.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-metrics-api.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jcommander-1.30.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-metastore.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-runner-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/HikariCP-2.6.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compress-1.9.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/guice-assistedinject-3.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-slf4j-impl-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ehcache-3.3.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop2-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/httpclient-4.5.3.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-mapreduce.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.inject-2.5.0-b32.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/objenesis-1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-common-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-server-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/velocity-1.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/janino-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-http.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/stax-api-1.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-lang3-3.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avatica-1.12.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/antlr-runtime-3.4.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-registry.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-server.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/leveldbjni-all-1.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-util-ajax-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jdo-api-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jline-2.12.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/taglibs-standard-impl-1.2.5.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-asl-1.9.13.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-io-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ant-launcher-1.9.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/audience-annotations-0.5.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/kite-hadoop-compatibility.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-server-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/javax.servlet.jsp-2.3.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/derby-10.14.1.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/findbugs-annotations-1.3.9-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-core-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/xz-1.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/geronimo-jcache_1.0_spec-1.0-alpha-1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-discovery-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-hadoop-compat.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-api-jdo-4.2.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/netty-3.10.6.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-xml-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/ivy-2.4.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/paranamer-2.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-compiler-2.7.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-client-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hbase-shaded-miscellaneous.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/avro-mapred-hadoop2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hadoop-yarn-server-applicationhistoryservice.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/websocket-servlet-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/datanucleus-rdbms-4.1.7.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/tephra-core-0.6.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service-rpc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jsr305-3.0.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-llap-common.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-mapper-asl-1.9.13-cloudera.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/json-20090211.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/validation-api-1.1.0.Final.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-service.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2-api-2.8.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jersey-guava-2.25.1.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-zookeeper-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jackson-core-2.9.8.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-format.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/log4j-1.2.17.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/oozie-sharelib-sqoop-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/dropwizard-metrics-hadoop-metrics2-reporter-0.1.2.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/jetty-annotations-9.3.25.v20180904.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/commons-io-2.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/parquet-jackson.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/twill-api-0.6.0-incubating.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/disruptor-3.3.6.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/sqoop/hive-orc.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie-5.1.0-cdh6.2.0.jar,hdfs://NameServiceOne/user/oozie/share/lib/lib_20190521153117/oozie/oozie-sharelib-oozie.jar dfs.client.slow.io.warning.threshold.ms : 30000 fs.s3a.multipart.size : 100M yarn.app.mapreduce.am.job.committer.commit-window : 10000 dfs.qjournal.new-epoch.timeout.ms : 120000 yarn.timeline-service.webapp.rest-csrf.enabled : false hadoop.proxyuser.flume.hosts : * dfs.namenode.edits.asynclogging : true yarn.timeline-service.reader.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineReaderImpl yarn.app.mapreduce.am.staging-dir.erasurecoding.enabled : false dfs.blockreport.incremental.intervalMsec : 0 dfs.datanode.network.counts.cache.max.size : 2147483647 dfs.namenode.https-address.NameServiceOne.namenode434 : node3:9871 yarn.timeline-service.writer.class : org.apache.hadoop.yarn.server.timelineservice.storage.HBaseTimelineWriterImpl mapreduce.ifile.readahead : true dfs.qjournal.get-journal-state.timeout.ms : 120000 yarn.timeline-service.entity-group-fs-store.summary-store : org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore dfs.client.socketcache.capacity : 16 fs.s3a.s3guard.ddb.table.create : false dfs.client.retry.policy.spec : 10000,6,60000,10 mapreduce.output.fileoutputformat.compress.codec : org.apache.hadoop.io.compress.DefaultCodec fs.s3a.socket.recv.buffer : 8192 dfs.datanode.fsdatasetcache.max.threads.per.volume : 4 dfs.namenode.reencrypt.batch.size : 1000 yarn.sharedcache.store.in-memory.initial-delay-mins : 10 mapreduce.jobhistory.webapp.address : masternode:19888 fs.adl.impl : org.apache.hadoop.fs.adl.AdlFileSystem fs.AbstractFileSystem.gs.impl : com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS mapreduce.task.userlog.limit.kb : 0 fs.s3a.connection.ssl.enabled : true yarn.router.rmadmin.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.rmadmin.DefaultRMAdminRequestInterceptor yarn.sharedcache.webapp.address : 0.0.0.0:8788 hadoop.fuse.connection.timeout : 300 dfs.http.client.retry.policy.spec : 10000,6,60000,10 yarn.resourcemanager.rm.container-allocation.expiry-interval-ms : 600000 ipc.server.max.connections : 0 yarn.app.mapreduce.am.resource.mb : 3072 hadoop.security.groups.cache.secs : 300 dfs.datanode.peer.stats.enabled : false dfs.replication : 3 mapreduce.shuffle.transfer.buffer.size : 131072 dfs.namenode.audit.log.async : false hadoop.security.group.mapping.ldap.directory.search.timeout : 10000 dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold : 10737418240 dfs.datanode.disk.check.timeout : 10m yarn.app.mapreduce.client-am.ipc.max-retries-on-timeouts : 3 fs.s3a.committer.threads : 8 dfs.checksum.combine.mode : MD5MD5CRC yarn.resourcemanager.nodemanager-graceful-decommission-timeout-secs : 3600 yarn.scheduler.maximum-allocation-vcores : 6 yarn.nodemanager.sleep-delay-before-sigkill.ms : 250 fs.AbstractFileSystem.abfs.impl : org.apache.hadoop.fs.azurebfs.Abfs mapreduce.job.acl-modify-job : fs.automatic.close : true fs.azure.sas.expiry.period : 90d dfs.qjm.operations.timeout : 60s hadoop.proxyuser.httpfs.hosts : * dfs.namenode.stale.datanode.minimum.interval : 3 dfs.namenode.reencrypt.edek.threads : 10 dfs.federation.router.store.membership.expiration : 300000 hadoop.security.groups.cache.background.reload.threads : 3 mapreduce.input.fileinputformat.list-status.num-threads : 1 hadoop.security.group.mapping.ldap.posix.attr.gid.name : gidNumber dfs.namenode.acls.enabled : false dfs.client.short.circuit.replica.stale.threshold.ms : 1800000 dfs.namenode.resource.du.reserved : 104857600 dfs.federation.router.connection.clean.ms : 10000 dfs.client.server-defaults.validity.period.ms : 3600000 dfs.federation.router.metrics.class : org.apache.hadoop.hdfs.server.federation.metrics.FederationRPCPerformanceMonitor mapreduce.shuffle.listen.queue.size : 128 mapreduce.jobhistory.intermediate-done-dir : ${yarn.app.mapreduce.am.staging-dir}/history/done_intermediate mapreduce.client.libjars.wildcard : true dfs.federation.router.cache.ttl : 60000 yarn.nodemanager.recovery.compaction-interval-secs : 3600 dfs.namenode.edits.noeditlogchannelflush : false mapreduce.reduce.shuffle.input.buffer.percent : 0.70 yarn.http.policy : HTTP_ONLY mapreduce.map.maxattempts : 4 dfs.namenode.audit.loggers : default io.serializations : org.apache.hadoop.io.serializer.WritableSerialization, org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization, org.apache.hadoop.io.serializer.avro.AvroReflectSerialization hadoop.security.groups.cache.warn.after.ms : 5000 dfs.client.write.byte-array-manager.count-reset-time-period-ms : 10000 yarn.nodemanager.webapp.rest-csrf.custom-header : X-XSRF-Header yarn.app.mapreduce.am.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH dfs.namenode.snapshot.capture.openfiles : true yarn.node-labels.fs-store.impl.class : org.apache.hadoop.yarn.nodelabels.FileSystemNodeLabelsStore hadoop.http.cross-origin.allowed-methods : GET,POST,HEAD dfs.qjournal.queued-edits.limit.mb : 10 mapreduce.jobhistory.webapp.rest-csrf.enabled : false dfs.http.policy : HTTP_ONLY dfs.balancer.max-size-to-move : 10737418240 dfs.datanode.sync.behind.writes.in.background : false hadoop.zk.acl : world:anyone:rwcda yarn.nodemanager.container.stderr.pattern : {*stderr*,*STDERR*} dfs.namenode.reencrypt.throttle.limit.updater.ratio : 1.0 mapreduce.cluster.local.dir : ${hadoop.tmp.dir}/mapred/local hadoop.kerberos.kinit.command : kinit dfs.namenode.secondary.https-address : 0.0.0.0:9869 dfs.namenode.metrics.logger.period.seconds : 600 dfs.block.access.token.lifetime : 600 dfs.ha.automatic-failover.enabled.NameServiceOne : true dfs.namenode.delegation.token.max-lifetime : 604800000 dfs.datanode.drop.cache.behind.writes : false dfs.mover.address : 0.0.0.0:0 dfs.block.placement.ec.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyRackFaultTolerant dfs.namenode.num.extra.edits.retained : 1000000 ipc.client.connect.max.retries.on.timeouts : 45 fs.viewfs.rename.strategy : SAME_MOUNTPOINT fs.client.resolve.topology.enabled : false hadoop.proxyuser.hive.hosts : * yarn.resourcemanager.node-labels.provider.fetch-interval-ms : 1800000 yarn.nodemanager.container-metrics.enable : true mapreduce.job.map.output.collector.class : org.apache.hadoop.mapred.MapTask$MapOutputBuffer fs.s3a.fast.upload.buffer : disk ha.health-monitor.connect-retry-interval.ms : 1000 dfs.namenode.edekcacheloader.initial.delay.ms : 3000 dfs.edit.log.transfer.bandwidthPerSec : 0 dfs.ha.tail-edits.in-progress : false dfs.federation.router.heartbeat.interval : 5000 ssl.client.truststore.reload.interval : 10000 dfs.client.datanode-restart.timeout : 30s io.mapfile.bloom.size : 1048576 hadoop.security.kms.client.authentication.retry-count : 1 dfs.client-write-packet-size : 65536 fs.ftp.data.connection.mode : ACTIVE_LOCAL_DATA_CONNECTION_MODE fs.swift.impl : org.apache.hadoop.fs.swift.snative.SwiftNativeFileSystem yarn.resourcemanager.webapp.rest-csrf.methods-to-ignore : GET,OPTIONS,HEAD mapreduce.job.max.map : -1 yarn.app.mapreduce.shuffle.log.backups : 0 ftp.blocksize : 67108864 dfs.namenode.kerberos.principal.pattern : * yarn.resourcemanager.scheduler.monitor.enable : false dfs.webhdfs.socket.connect-timeout : 60s dfs.namenode.replication.max-streams : 2 nfs.allow.insecure.ports : true yarn.sharedcache.nm.uploader.thread-count : 20 dfs.federation.router.admin.enable : true yarn.app.mapreduce.client.job.retry-interval : 2000 yarn.scheduler.configuration.store.max-logs : 1000 hadoop.security.authorization : false yarn.timeline-service.version : 1.0f yarn.am.liveness-monitor.expiry-interval-ms : 600000 fs.har.impl.disable.cache : true hadoop.proxyuser.hdfs.hosts : * mapreduce.job.reduce.slowstart.completedmaps : 0.8 yarn.timeline-service.leveldb-timeline-store.path : ${hadoop.tmp.dir}/yarn/timeline dfs.namenode.upgrade.domain.factor : ${dfs.replication} mapreduce.jobhistory.minicluster.fixed.ports : false mapreduce.application.classpath : $HADOOP_CLIENT_CONF_DIR,$PWD/mr-framework/*,$MR2_CLASSPATH yarn.resourcemanager.delegation.token.max-lifetime : 604800000 yarn.resourcemanager.ha.automatic-failover.enabled : true mapreduce.reduce.java.opts : -Djava.net.preferIPv4Stack=true dfs.datanode.socket.write.timeout : 480000 dfs.namenode.accesstime.precision : 3600000 dfs.namenode.redundancy.considerLoad.factor : 2.0 yarn.resourcemanager.store.class : org.apache.hadoop.yarn.server.resourcemanager.recovery.FileSystemRMStateStore io.mapfile.bloom.error.rate : 0.005 yarn.nodemanager.webapp.rest-csrf.enabled : false yarn.timeline-service.leveldb-state-store.path : ${hadoop.tmp.dir}/yarn/timeline hadoop.proxyuser.hive.groups : * dfs.federation.router.rpc-address : 0.0.0.0:8888 fs.s3a.committer.staging.unique-filenames : true dfs.namenode.support.allow.format : true yarn.scheduler.configuration.zk-store.parent-path : /confstore dfs.content-summary.limit : 5000 yarn.timeline-service.writer.flush-interval-seconds : 60 yarn.nodemanager.container-executor.class : org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor dfs.namenode.posix.acl.inheritance.enabled : true dfs.datanode.outliers.report.interval : 30m hadoop.security.kms.client.encrypted.key.cache.low-watermark : 0.3f dfs.namenode.top.enabled : true yarn.app.mapreduce.shuffle.log.separate : true hadoop.user.group.static.mapping.overrides : dr.who=; dfs.federation.router.http-address : 0.0.0.0:50071 fs.s3a.retry.throttle.interval : 1000ms yarn.nodemanager.amrmproxy.address : 0.0.0.0:8049 mapreduce.jobhistory.webapp.rest-csrf.custom-header : X-XSRF-Header yarn.webapp.xfs-filter.enabled : true dfs.client.cached.conn.retry : 3 dfs.client.key.provider.cache.expiry : 864000000 dfs.namenode.path.based.cache.refresh.interval.ms : 30000 yarn.nodemanager.collector-service.thread-count : 5 dfs.block.replicator.classname : org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicyDefault dfs.namenode.fs-limits.max-directory-items : 1048576 dfs.ha.log-roll.period : 120s yarn.nodemanager.runtime.linux.docker.capabilities : CHOWN,DAC_OVERRIDE,FSETID,FOWNER,MKNOD,NET_RAW,SETGID,SETUID,SETFCAP,SETPCAP,NET_BIND_SERVICE,SYS_CHROOT,KILL,AUDIT_WRITE yarn.nodemanager.distributed-scheduling.enabled : false ipc.client.fallback-to-simple-auth-allowed : false yarn.minicluster.fixed.ports : false yarn.nodemanager.remote-app-log-dir : /tmp/logs yarn.timeline-service.entity-group-fs-store.scan-interval-seconds : 60 dfs.xframe.enabled : true yarn.nodemanager.resource.percentage-physical-cpu-limit : 100 mapreduce.job.tags : oozie-564a124254f1fd53cb03553181f7e603 dfs.namenode.fs-limits.max-xattr-size : 16384 dfs.datanode.http.address : 0.0.0.0:9864 dfs.namenode.blocks.per.postponedblocks.rescan : 10000 fs.s3a.s3guard.cli.prune.age : 86400000 dfs.web.authentication.filter : org.apache.hadoop.hdfs.web.AuthFilter dfs.namenode.maintenance.replication.min : 1 hadoop.jetty.logs.serve.aliases : true dfs.webhdfs.ugi.expire.after.access : 600000 dfs.namenode.max.op.size : 52428800 mapreduce.jobhistory.admin.acl : * mapreduce.job.reducer.unconditional-preempt.delay.sec : 300 yarn.app.mapreduce.am.hard-kill-timeout-ms : 10000 yarn.resourcemanager.display.per-user-apps : false yarn.resourcemanager.node-removal-untracked.timeout-ms : 60000 yarn.resourcemanager.webapp.address : masternode:8088 mapreduce.jobhistory.recovery.enable : false yarn.sharedcache.store.in-memory.check-period-mins : 720 dfs.client.test.drop.namenode.response.number : 0 fs.df.interval : 60000 fs.s3a.assumed.role.session.duration : 30m mapreduce.job.cache.limit.max-single-resource-mb : 0 yarn.timeline-service.enabled : false dfs.disk.balancer.block.tolerance.percent : 10 dfs.webhdfs.netty.high.watermark : 65535 mapreduce.task.profile : false hadoop.http.cross-origin.allowed-headers : X-Requested-With,Content-Type,Accept,Origin yarn.router.webapp.address : 0.0.0.0:8089 dfs.datanode.balance.max.concurrent.moves : 50 yarn.nodemanager.hostname : 0.0.0.0 mapreduce.task.exit.timeout : 60000 yarn.resourcemanager.nm-container-queuing.max-queue-length : 15 mapreduce.job.token.tracking.ids.enabled : false yarn.scheduler.increment-allocation-mb : 512 fs.s3a.assumed.role.credentials.provider : org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider fs.azure.authorization.caching.enable : true hadoop.security.kms.client.failover.sleep.max.millis : 2000 dfs.client.mmap.retry.timeout.ms : 300000 yarn.resourcemanager.webapp.rest-csrf.custom-header : X-XSRF-Header yarn.resourcemanager.nm-container-queuing.max-queue-wait-time-ms : 100 mapreduce.jobhistory.move.thread-count : 3 dfs.permissions.enabled : true fs.AbstractFileSystem.hdfs.impl : org.apache.hadoop.fs.Hdfs yarn.nodemanager.container-localizer.log.level : INFO hadoop.http.filter.initializers : org.apache.hadoop.http.lib.StaticUserWebFilter yarn.timeline-service.http-authentication.simple.anonymous.allowed : true yarn.nodemanager.runtime.linux.docker.allowed-container-networks : host,none,bridge dfs.qjournal.accept-recovery.timeout.ms : 120000 yarn.sharedcache.client-server.thread-count : 50 fs.s3a.s3guard.ddb.max.retries : 9 fs.s3a.committer.magic.enabled : false yarn.resourcemanager.resource-tracker.address : masternode:8031 mapreduce.jobhistory.jobname.limit : 50 dfs.domain.socket.path : /var/run/hdfs-sockets/dn dfs.namenode.decommission.blocks.per.interval : 500000 dfs.qjournal.write-txns.timeout.ms : 20000 rpc.metrics.quantile.enable : false yarn.federation.subcluster-resolver.class : org.apache.hadoop.yarn.server.federation.resolver.DefaultSubClusterResolverImpl dfs.namenode.read-lock-reporting-threshold-ms : 5000 mapreduce.task.timeout : 600000 yarn.nodemanager.resource.memory-mb : -1 dfs.datanode.failed.volumes.tolerated : 0 yarn.nodemanager.disk-health-checker.min-healthy-disks : 0.25 mapreduce.framework.name : yarn mapreduce.fileoutputcommitter.algorithm.version : 2 yarn.router.clientrm.interceptor-class.pipeline : org.apache.hadoop.yarn.server.router.clientrm.DefaultClientRequestInterceptor yarn.sharedcache.nested-level : 3 fs.s3a.connection.timeout : 200000 hadoop.caller.context.signature.max.size : 40 hadoop.security.dns.log-slow-lookups.enabled : false mapreduce.jobhistory.webapp.https.address : masternode:19890 file.client-write-packet-size : 65536 fs.s3a.s3guard.ddb.table.capacity.read : 500 ipc.client.ping : true hadoop.proxyuser.oozie.hosts : * yarn.resourcemanager.delayed.delegation-token.removal-interval-ms : 30000 dfs.client.failover.max.attempts : 15 dfs.balancer.max-no-move-interval : 60000 yarn.nodemanager.opportunistic-containers-use-pause-for-preemption : false yarn.nodemanager.webapp.cross-origin.enabled : false mapreduce.job.encrypted-intermediate-data : false dfs.client.read.shortcircuit.streams.cache.expiry.ms : 300000 yarn.minicluster.control-resource-monitoring : false dfs.disk.balancer.enabled : false dfs.webhdfs.oauth2.enabled : false yarn.nodemanager.health-checker.script.timeout-ms : 1200000 yarn.resourcemanager.fs.state-store.num-retries : 0 hadoop.ssl.require.client.cert : false mapreduce.jobhistory.keytab : /etc/security/keytab/jhs.service.keytab hadoop.security.uid.cache.secs : 14400 yarn.resourcemanager.ha.automatic-failover.zk-base-path : /yarn-leader-election yarn.intermediate-data-encryption.enable : false mapreduce.job.speculative.speculative-cap-running-tasks : 0.1 dfs.datanode.block.id.layout.upgrade.threads : 12 dfs.client.context : default yarn.system-metrics-publisher.enabled : false dfs.namenode.delegation.token.renew-interval : 86400000 yarn.timeline-service.entity-group-fs-store.app-cache-size : 10 fs.AbstractFileSystem.s3a.impl : org.apache.hadoop.fs.s3a.S3A mapreduce.job.redacted-properties : fs.s3a.access.key,fs.s3a.secret.key,fs.adl.oauth2.credential,dfs.adls.oauth2.credential,fs.azure.account.oauth2.client.secret yarn.client.load.resource-types.from-server : false ipc.client.tcpnodelay : true hadoop.proxyuser.httpfs.groups : * yarn.resourcemanager.metrics.runtime.buckets : 60,300,1440 dfs.blockreport.intervalMsec : 21600000 dfs.datanode.oob.timeout-ms : 1500,0,0,0 yarn.client.application-client-protocol.poll-timeout-ms : -1 zlib.compress.level : DEFAULT_COMPRESSION mapreduce.job.sharedcache.mode : disabled io.map.index.skip : 0 mapreduce.job.hdfs-servers : ${fs.defaultFS} mapreduce.map.output.compress : true hadoop.security.kms.client.encrypted.key.cache.num.refill.threads : 2 dfs.namenode.edekcacheloader.interval.ms : 1000 mapreduce.task.merge.progress.records : 10000 yarn.nodemanager.aux-services.mapreduce_shuffle.class : org.apache.hadoop.mapred.ShuffleHandler dfs.namenode.missing.checkpoint.periods.before.shutdown : 3 tfile.fs.output.buffer.size : 262144 dfs.client.failover.connection.retries : 0 fs.du.interval : 600000 dfs.edit.log.transfer.timeout : 30000 dfs.namenode.top.window.num.buckets : 10 dfs.data.transfer.server.tcpnodelay : true hadoop.zk.retry-interval-ms : 1000 yarn.sharedcache.uploader.server.address : 0.0.0.0:8046 dfs.http.client.failover.max.attempts : 15 fs.s3a.socket.send.buffer : 8192 dfs.client.block.write.locateFollowingBlock.retries : 7 hadoop.registry.zk.quorum : localhost:2181 mapreduce.jvm.system-properties-to-log : os.name,os.version,java.home,java.runtime.version,java.vendor,java.version,java.vm.name,java.class.path,java.io.tmpdir,user.dir,user.name hadoop.http.cross-origin.allowed-origins : * dfs.namenode.enable.retrycache : true dfs.datanode.du.reserved : 0 hadoop.registry.system.acls : sasl:yarn@, sasl:mapred@, sasl:hdfs@ yarn.nodemanager.resource-plugins.gpu.docker-plugin.nvidia-docker-v1.endpoint : http://localhost:3476/v1.0/docker/cli mapreduce.job.encrypted-intermediate-data.buffer.kb : 128 dfs.data.transfer.client.tcpnodelay : true yarn.resourcemanager.webapp.xfs-filter.xframe-options : SAMEORIGIN mapreduce.admin.user.env : LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:$JAVA_LIBRARY_PATH mapreduce.task.profile.reduce.params : ${mapreduce.task.profile.params} mapreduce.reduce.memory.mb : 0 hadoop.caller.context.enabled : false hadoop.http.authentication.kerberos.principal : HTTP/_HOST@LOCALHOST yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb : 0 dfs.qjournal.prepare-recovery.timeout.ms : 120000 dfs.datanode.transferTo.allowed : true oozie.action.rootlogger.log.level : INFO hadoop.security.sensitive-config-keys : secret$ password$ ssl.keystore.pass$ fs.s3.*[Ss]ecret.?[Kk]ey fs.s3a.*.server-side-encryption.key fs.azure.account.key.* credential$ oauth.*secret oauth.*password oauth.*token hadoop.security.sensitive-config-keys mapreduce.client.completion.pollinterval : 5000 dfs.namenode.name.dir.restore : false dfs.namenode.full.block.report.lease.length.ms : 300000 dfs.namenode.secondary.http-address : 0.0.0.0:9868 hadoop.http.logs.enabled : true hadoop.security.group.mapping.ldap.read.timeout.ms : 60000 yarn.resourcemanager.max-log-aggregation-diagnostics-in-memory : 10 dfs.namenode.delegation.token.always-use : false yarn.resourcemanager.webapp.https.address : masternode:8090 fs.s3a.retry.throttle.limit : ${fs.s3a.attempts.maximum} dfs.client.read.striped.threadpool.size : 18 mapreduce.job.cache.limit.max-resources : 0 hadoop.proxyuser.HTTP.groups : * -------------------- Setting up log4j2 log4j2 configuration file created at /yarn/nm/usercache/admin/appcache/application_1560674082717_0001/container_1560674082717_0001_01_000001/sqoop-log4j2.xml Sqoop command arguments : import \ --connect 'jdbc:sqlserver://myServer;database=myDB' \ --username myUsername --password ******** \ --table category -m 1 --check-column LastEditOn \ --merge-key 'Reference ID' \ --incremental lastmodified \ --compression-codec=snappy \ --target-dir /user/hive/warehouse/myDB.db/category \ --hive-table category \ --map-column-hive LastEditOn=timestamp,CreatedOn=timestamp \ --fields-terminated-by '\001' --fields-terminated-by '\n' Fetching child yarn jobs tag id : oozie-564a124254f1fd53cb03553181f7e603 No child applications found ================================================================= >>> Invoking Sqoop command line now >>> <<< Invocation of Sqoop command completed <<< No child hadoop job is executed. java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.oozie.action.hadoop.LauncherAM.runActionMain(LauncherAM.java:410) at org.apache.oozie.action.hadoop.LauncherAM.access$300(LauncherAM.java:55) at org.apache.oozie.action.hadoop.LauncherAM$2.run(LauncherAM.java:223) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875) at org.apache.oozie.action.hadoop.LauncherAM.run(LauncherAM.java:217) at org.apache.oozie.action.hadoop.LauncherAM$1.run(LauncherAM.java:153) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875) at org.apache.oozie.action.hadoop.LauncherAM.main(LauncherAM.java:141) Caused by: java.lang.SecurityException: Intercepted System.exit(1) at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57) at java.lang.Runtime.exit(Runtime.java:107) at java.lang.System.exit(System.java:971) at org.apache.sqoop.Sqoop.main(Sqoop.java:252) at org.apache.oozie.action.hadoop.SqoopMain.runSqoopJob(SqoopMain.java:214) at org.apache.oozie.action.hadoop.SqoopMain.run(SqoopMain.java:199) at org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:104) at org.apache.oozie.action.hadoop.SqoopMain.main(SqoopMain.java:51) ... 16 more Intercepting System.exit(1) Failing Oozie Launcher, Main Class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1] Oozie Launcher, uploading action data to HDFS sequence file: hdfs://NameServiceOne/user/admin/oozie-oozi/0000000-190616123600049-oozie-oozi-W/sqoop-c9e7--sqoop/action-data.seq 12:41:09.783 [main] INFO org.apache.hadoop.io.compress.CodecPool - Got brand-new compressor [.deflate] Stopping AM 12:41:09.983 [main] INFO org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl - Waiting for application to be successfully unregistered. Callback notification attempts left 0 Callback notification trying http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED Callback notification to http://masternode:11000/oozie/callback?id=0000000-190616123600049-oozie-oozi-W@sqoop-c9e7&status=FAILED succeeded Callback notification succeeded
Created 06-16-2019 05:35 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thanks for sharing the output. From the error, looks like that Sqoop did not parse the command correctly. And I noticed below parameter:
--merge-key
'Reference
ID'
Noticed the line break between Reference and ID? I am not sure if it is formatting issue in the forum or it was like this in your Sqoop job setup. Can you please double check on this? And make sure there is no new line between the two words?
Cheers
Eric
Created 06-19-2019 01:03 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I am sorry but there is nothing wrong with the syntax, as if I run it on the terminal it completes successfully.
I have a doubt regarding security, and that is because of the following line in the log
aused by: java.lang.SecurityException: Intercepted System.exit(1) at org.apache.oozie.action.hadoop.security.LauncherSecurityManager.checkExit(LauncherSecurityManager.java:57)
I would like to note also that this is the first try to run sqoop script on Hue after a fresh installation of cdh 6.2... so I am afraid there is something that I've missed in the configuration, but I really can't find it 😞
Created 06-19-2019 04:04 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Please also share the screenshot of the Sqoop job configuration in Hue for review.
Cheers
Eric
Created 06-19-2019 11:31 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I hope this is what you want for the workflow:
<workflow-app name="Batch job for query-sqoop1" xmlns="uri:oozie:workflow:0.5"> <start to="sqoop-fde5"/> <kill name="Kill"> <message>Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message> </kill> <action name="sqoop-fde5"> <sqoop xmlns="uri:oozie:sqoop-action:0.2"> <job-tracker>${jobTracker}</job-tracker> <name-node>${nameNode}</name-node> <command>import \ --connect 'jdbc:sqlserver://11.11.11.11;database=DBXYZ' \ --username theUser --password thePassword \ --table category -m 1 --check-column LastEditOn \ --merge-key 'Reference ID' \ --incremental lastmodified \ --compression-codec=snappy \ --target-dir /user/hive/warehouse/dwh_db_atlas_jrtf.db/category \ --hive-table category \ --map-column-hive LastEditOn=timestamp,CreatedOn=timestamp \ --fields-terminated-by '\001' --fields-terminated-by '\n'</command> </sqoop> <ok to="End"/> <error to="Kill"/> </action> <end name="End"/> </workflow-app>
For the job configuration, I am really not sure where to find it. The one I can reach is a lot of scrolls that can't be taken screenshot in anyway.... so would you please give me the path to the job configuration ?
