Member since
11-18-2014
20
Posts
10
Kudos Received
0
Solutions
12-16-2015
09:53 PM
1 Kudo
I tried, but nothing changed. Where should I configure the collector interval. I don't need a "real-time" monitoring. Maybe 1 measurement per minute is enough.
... View more
12-16-2015
07:01 PM
1 Kudo
Here is my config files for ambari-metrics-collector ambari-metrics-collector.tar.gz
... View more
12-16-2015
06:41 PM
1 Kudo
You did not answered my question completly. I'm the only user on cluster. There's no reason for 100% cpu.
... View more
12-16-2015
06:20 PM
1 Kudo
Indeed. I found this "internal HBase" version. Each node of cluster has 32 GB of RAM and 4 vCores. All of them are virtualized on top of very good hardware.
... View more
12-16-2015
06:19 PM
1 Kudo
I added this line to ambari.properties and restarted ambari-server, but the 100% cpu behavior stands. Se the screenshot in comment below.
... View more
12-16-2015
04:32 PM
1 Kudo
I have a HDP 2.3.0 cluster with 4 nodes. I noticed this process was consuming 100% cpu on my NameNode: ams 5386 223 5.0 3596120 1666616 ? Sl 13:50 46:56 /usr/jdk64/jdk1.8.0_40/bin/java -Dproc_master -XX:OnOutOfMemoryError=kill -9 %p -Xmx1536m -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/ambari-metrics-collector/hs_err_pid%p.log -Djava.io.tmpdir=/opt/var/lib/ambari-metrics-collector/hbase-tmp -Djava.library.path=/usr/lib/ams-hbase/lib/hadoop-native/ -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/ambari-metrics-collector/gc.log-201512161350 -Xms1536m -Xmx1536m -Xmn256m -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly -Dhbase.log.dir=/var/log/ambari-metrics-collector -Dhbase.log.file=hbase-ams-master-NPAA1809.petrobras.biz.log -Dhbase.home.dir=/usr/lib/ams-hbase/bin/.. -Dhbase.id.str=ams -Dhbase.root.logger=INFO,RFA -Dhbase.security.logger=INFO,RFAS org.apache.hadoop.hbase.master.HMaster start But my cluster does not have Hbase installed. Then I simply killed this process, but Ambari metrics went down together. After I restarted Ambari metrics, this process continues trying to executing and always consuming a lot of CPU. Here is a print of Ambari Metrics durint time of my actions. How can I configure AMS to stop trying to monitor HBase?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache HBase
12-15-2015
09:27 AM
From my own experience, Spark runs much faster in standalone mode. I tried a variety of configurations on Yarn, but I can't get same performance. I'll try to upgrade. Is there a guide to standalone mode?
... View more
12-14-2015
03:06 PM
1 Kudo
I'm trying to calculate an averange of randomForest predictions in Spark 1.3.1, since the predicted probability of all trees is available only in 1.5.0. The best I could do until now is using the function below: def calculaProbs(dados, modelRF):
trees = modelRF._java_model.trees()
nTrees = modelRF.numTrees()
nPontos = dados.count()
predictions = np.zeros(nPontos)
for i in range(nTrees):
dtm = DecisionTreeModel(trees[i])
predictions+= dtm.predict(dados.map(lambda x: x.features)).collect()
predictions = predictions/nTrees
return predictions This code is running very slow, as expected, since I'm collecting (collect()) predictions from each Tree and adding them up in Driver. I cannot put the dtm.predit() inside a Map operation in this version of Spark. Here is the Note from documentation: "Note: In Python, predict cannot currently be used within an RDD transformation or action. Call predict directly on the RDD instead." Any Idea to improve performance? How can I add values from 2 RDDs without collecting their values to a vector?
... View more
Labels:
- Labels:
-
Apache Spark
12-10-2015
08:28 PM
before calling this routine, I introduced the code bellow and exec time reduced to 1m8s. 3x improvement. rawTrainData = rawTrainData.repartition(8)
rawTrainData.cache() But introducing numPartitions=15 inside distinct method does not affect the result. I'm running Spark 1.3.1 into standalone mode (spark://host:7077) with 12 cores and 20 GB per node allocated to Spark. The hardware is virtual, but I know it`s a top hardware. The cluster has 4 nodes (3 spark workers)
... View more