Member since
02-01-2019
650
Posts
143
Kudos Received
117
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2822 | 04-01-2019 09:53 AM | |
1463 | 04-01-2019 09:34 AM | |
6927 | 01-28-2019 03:50 PM | |
1577 | 11-08-2018 09:26 AM | |
3802 | 11-08-2018 08:55 AM |
09-09-2018
09:17 AM
1 Kudo
@Daniel
Zafar
Apache Tez replaces MapReduce as the default Hive execution engine in HDP 3.0. MapReduce is no longer supported. You may want to check what is the actual issue with Tez and fix it. Ref: https://docs.hortonworks.com/HDPDocuments/HDP3/HDP-3.0.0/hive-overview/content/hive-apache-hive-3-architecturural-overview.html
... View more
09-07-2018
12:36 PM
@Kumar Veerappan Sizing HiveServer2 Heap Memory
The following are general recommendations for sizing heap memory of a HiveServer2 instance:
1 to 20 concurrent executing queries: Set to 6 GB heap size.
21 to 40 concurrent executing queries: Set to 12 GB heap size.
More than 40 concurrent executing queries: Create a new HiveServer2 instance. See Multiple HiveServer2Instances for Different Workloads for how to add a HiveServer2 instance.
Please refer the official doc : https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.5/bk_hive-performance-tuning/content/ch_connectivity-admission-control.html#guidelines-hiveserver2-heaps
... View more
09-07-2018
11:20 AM
1 Kudo
@Ronnie 10 Ambari will by default pick up the mount points and configure them to appropriate services. eg: For HDFS ambari configures dfs.datanode.data.dir and dfs.namenode.data.dir with all the mount points. Do when you start using HDFS you should see data insode your /data0 /data1 /data2 and so on. Hope this helps.
... View more
09-06-2018
01:09 PM
@Michael Bronson By default Spark2 has log level as WARN. Set it to INFO to get more context on what is going on in the driver and executor. More over the log will be locally available in Nodemanager when the container is still running. The easiest way is to go to spark UI (yarn application master UI) -> click on executors tab -> Here you should see stderr and stdout corresponding to driver and executors. Regarding the WARN on heartbeat , we'd need to check what driver is doing at that point. I think you already have asked another question with more details on driver and executor.
... View more
09-06-2018
09:11 AM
@Michael Bronson Spark will not log anything in Datanode machines(where executors/containers are running) at /var/log/spark2. Spark app is like any other yarn application. When the application is running the logs will be stored in the container home directory and then it will be moved to hdfs post log aggregation(which can be extracted by yarn logs command). Hope this helps.
... View more
09-06-2018
08:25 AM
@Michael Bronson, You will see Spark thrift server and Spark History Server in /var/log/spark2. The above log4j i proposed is for spark applications (which will not be stored in /var/log/spark2 rather you should use yarn logs command and extract the log). What is it you want to enable DEBUG logging for? spark application , Spark Thrift server or Spark History Server?
... View more
09-05-2018
04:23 PM
@Michael Bronson Use below : # Set everything to be logged to the console log4j.rootCategory=DBEUG, console log4j.appender.console=org.apache.log4j.ConsoleAppender log4j.appender.console.target=System.err log4j.appender.console.layout=org.apache.log4j.PatternLayout log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss}%p %c{1}:%m%n # Settings to quiet third party logs that are too verbose log4j.logger.org.eclipse.jetty=WARN log4j.logger.org.eclipse.jetty.util.component.AbstractLifeCycle=ERROR log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO log4j.logger.org.apache.spark.metrics.MetricsConfig=DEBUG log4j.logger.org.apache.spark.deploy.yarn.Client=DEBUG
... View more
08-30-2018
02:58 PM
@David Hoyle
The code structure has changed since this article was written. 1) checkout trunk 2) brew install protobuf250 (protobuf is needed to build hadoop) 3) Build using : mvn clean package -Phdds -Pdist -Dtar -DskipShade -DskipTests -Dmaven.javadoc.skip=true edit: updated the proto version
... View more
08-27-2018
06:47 AM
This Blog series would give you the complete overview of Hive+Druid : https://hortonworks.com/blog/apache-hive-druid-part-1-3/
... View more
08-21-2018
04:42 AM
Not Yet...
... View more