Member since
04-07-2017
10
Posts
0
Kudos Received
0
Solutions
07-12-2017
08:24 AM
My issue here ended up being brought by Amazon elastic load balancer timing out after 60 seconds.
... View more
06-09-2017
04:14 PM
All Hive sessions (via Hive JDBC on Squirrel SQL) are closed after 60 seconds of idle time regardless of configuration options: hive.server2.session.check.interval = 0 hive.server2.idle.operation.timeout = 0 hive.server2.idle.session.timeout = 0 hive.server2.idle.session.timeout_check_operation = false Why are these options effectively ignored and how can it be fixed? Hive version is 1.1.0.
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Spark
-
Apache YARN
05-25-2017
10:56 AM
@mbigelowFair point. I did a write up specifically for this forum, but apparently my session had expired so hitting post lost it all, so I didn't want to re-write it; my bad. I did try setting `spark.dynamicAllocation.enabled` to false, which still didn't change the decay issue. I even upped `spark.dynamicAllocation.executorIdleTimeout` to 5 minutes (from 1 minute) in case that was the problem, but it didn't seem to have an effect. My main theory right now is that because the data I'm accessing is on HDFS and is minimally replicated, _maybe_ the executors are dropped because they supposedly don't think they can work... I'm going to try building an external table in a similar vein to my HDFS one using my same data in S3 parquet. Here are the details from SO: Whether I use dynamic allocation or explicitly specify executors (16) and executor cores (8), I have been losing executors even though the tasks outstanding are well beyond the current number of executors. For example, I have a job (Spark SQL) running with over 27,000 tasks and 14,000 of them were complete, but executors "decayed" from 128 down to as few as 16 with thousands of tasks still outstanding. The log doesn't note any errors/exceptions precipitating these lost executors. It is a Cloudera CDH 5.10 cluster running on AWS EC2 instances with 136 CPU cores and Spark 2.1.0 (from Cloudera). 17/05/23 18:54:17 INFO yarn.YarnAllocator: Driver requested a total number of 91 executor(s).
17/05/23 18:54:17 INFO yarn.YarnAllocator: Canceling requests for 1 executor container(s) to have a new desired total 91 executors. It's a slow decay where every minute or so more executors are removed. Some potentially relevant configuration options: spark.dynamicAllocation.maxExecutors = 136
spark.dynamicAllocation.minExecutors = 1
spark.dynamicAllocation.initialExecutors = 1
yarn.nodemanager.resource.cpu-vcores = 8
yarn.scheduler.minimum-allocation-vcores = 1
yarn.scheduler.increment-allocation-vcores = 1
yarn.scheduler.maximum-allocation-vcores = 8 Why are the executors decaying away and how can I prevent it?
... View more
05-24-2017
05:11 PM
I asked this question on SO, but Cloudera forums may be a better place... https://stackoverflow.com/questions/44143230/losing-spark-executors-with-many-tasks-outstanding Why am I losing my executors and how can I stop it?
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache YARN
04-25-2017
12:15 PM
Thanks @aarman; that worked wonderfully!
... View more
04-21-2017
10:43 AM
@task
def install_jdk8_centos():
sudo("yum install -y wget")
sudo('cd /opt/ && tar xzf jdk-8u121-linux-x64.tar.gz && cd /opt/jdk1.8.0_121/')
sudo("alternatives --install /usr/bin/java java /opt/jdk1.8.0_121/bin/java 2")
sudo("alternatives --set java /opt/jdk1.8.0_121/bin/java")
sudo("alternatives --install /usr/bin/jar jar /opt/jdk1.8.0_121/bin/jar 2")
sudo("alternatives --install /usr/bin/javac javac /opt/jdk1.8.0_121/bin/javac 2")
sudo("alternatives --set jar /opt/jdk1.8.0_121/bin/jar")
sudo("alternatives --set javac /opt/jdk1.8.0_121/bin/javac")
sudo("export JAVA_HOME=/opt/jdk1.8.0_121")
sudo("export JRE_HOME=/opt/jdk1.8.0_121/jre")
sudo("export PATH=$PATH:/opt/jdk1.8.0_121/bin:/opt/jdk1.8.0_121/jre/bin")
sudo("echo JAVA_HOME=/opt/jdk1.8.0_121 | tee -a /etc/environment")
sudo("echo JRE_HOME=/opt/jdk1.8.0_121/jre | tee -a /etc/environment")
sudo("echo PATH=$PATH:/opt/jdk1.8.0_121/bin:/opt/jdk1.8.0_121/jre/bin | tee -a /etc/environment") With that style of fabfile fabric script and setting the java home directory in CM, the infamous unsupported major.minor problem was vanquished for me. Source of exquisite instructions: https://tecadmin.net/install-java-8-on-centos-rhel-and-fedora/
... View more
04-20-2017
03:20 PM
Hey @Bill Havanki I edited the files and added these lines: ec2.virtualizationmappings.properties hvm=i3.large,\ i3.xlarge,\ i3.2xlarge,\ i3.4xlarge,\ i3.8xlarge,\ i3.16xlarge,\ ec2.ephemeraldevicemappings.properties: i3.large=1 i3.xlarge=1 i3.2xlarge=1 i3.4xlarge=2 i3.8xlarge=4 i3.16xlarge=8 Unfortunately, while the provisioning of the instances was successful, the NVMe drives were NOT mounted. This is likely because they are at non-default locations like "/dev/nvme0n1"
... View more
04-07-2017
02:05 PM
Thanks Bill! Oddly, I had no problem launching an i3.8xlarge with that same AMI (ami-6d1c2007)
... View more
04-07-2017
01:03 PM
When will the new i3 instances be supported in Cloudera Director? When I tried to launch a cluster with the i3 instances, it threw a generic error about how that instance type is not supported by the AMI I selected. I then made up a non-existent instance type and got the same error.
... View more
Labels:
- Labels:
-
Cloudera Manager