Member since
09-24-2015
816
Posts
488
Kudos Received
189
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2572 | 12-25-2018 10:42 PM | |
11842 | 10-09-2018 03:52 AM | |
4134 | 02-23-2018 11:46 PM | |
1796 | 09-02-2017 01:49 AM | |
2114 | 06-21-2017 12:06 AM |
04-29-2016
09:15 AM
1 Kudo
You can locate them through ambari. When you (re)start a service you can click on the operations > operation > tasks and inspect the commands: If you look closely the script being executed for restarting the nodemanager is at 08:53:13,592. The script is located in /usr/hdp/current/hadoop-yarn-nodemanager/sbin/yarn-daemon.sh. This file is shipped with the distribution. Before executing this file users are created and config is pushed. The preparation of these steps happen on the AmbariServer. You can search for the python scripts. For example the nodemanager in the /var/lib/ambari-server/resources/common-services/YARN/2.1.0.2.0/package/scripts/. If you change one these files, don't forget to restart the ambari-server, because the files are cached. After an ambari-server upgrade these changes will be overridden reverted. Hope this helps.
... View more
04-18-2016
02:45 PM
Brilliant @Predrag Minovic , that solved the issue. Thanks !
... View more
04-16-2016
06:29 PM
Thank you minovic. You are very helpful
... View more
04-15-2016
06:09 AM
1 Kudo
Thanks Laurence Da Luz . It helped me
... View more
04-15-2016
05:56 AM
Ah, okay, I understand now. Steps are already given in the Ambari Upgrade guide, though some of them have many details, and talk about other OS like Suse and Ubuntu, and talk about Kerberos yes/no, Ganglia yes/no and so on. You can extract only steps applicable to your cluster from Chapters 2, 3, and 4 of the guide: Chapter 2. Getting Ready to Upgrade Ambari and HDP Chapter 3. Upgrading Ambari Chapter 4. Upgrading HDP and then as you proceed with the upgrade you can mark completed steps, and for each step you can write down any comments/observations to refer to them later.
... View more
04-12-2016
09:32 AM
Hi @jeff, I have the same problem after upgrade Amabri 2.2.0.0 to 2.2.1.1 directly. The result of the query is: ambari=> select repo_version_id,stack_id,version,display_name from repo_version;
repo_version_id | stack_id | version | display_name
-----------------+----------+--------------+------------------
1 | 4 | 2.3.4.0-3485 | HDP-2.3.4.0-3485
51 | 51 | 2.4.0.0 | HDP-2.4.0.0
(2 righe)
... View more
12-20-2017
08:54 AM
hi all, i'm still getting "PutKafka Error : failed while waiting for acks from kafka" error though i see some values under "In" & "Read". but i see attached consumer console error. Also,i see something strange here. When i remove out Kafka from "maintenance mode" on Ambari and start the kafka broker, it gets stopped by itself after a while. please help me on this & find supporting attachments. @jmedel @Predrag Minovic @Artem Ervits Note: I'm using HDP 2.6 Regards, Akshay putkafka-processor-nifi-properties.png
... View more
04-10-2019
05:20 PM
@Benjamin Leonhardi could you please tell me how to actually increase the hash cache? I agree 100MB is not much. I'm running an HDP 2.6.5 cluster with HBase 1.1.2 and Phoenix enabled. I couldn't find the property phoenix.query.maxServerCacheBytes defined in hbase-site.xml config file, so I tried adding it in the "Custom hbase-site" section in Ambari and setting it to 500Mb as shown in the picture below: Now I can see the variable defined in hbase-site.xml, but I am still getting the following error "...maximum allowed size (104857600 bytes)" even though I have changed that setting to 500MB). Any ideas or suggestions? Thanks in advance!
... View more
09-14-2016
10:08 AM
@suno bella I had same issue , the problem was I added my mount command of shared folder in /etc/profile.d/mount.sh
afte I moved it to /etc/init.d/ the issue was resloved , I think the problem was as I logged in HDFS it was calling the mount.sh again . ,
... View more
04-03-2016
05:23 AM
@suno bella, no problems, if you found my answers useful please consider to accept and/or up-vote my first answer above. Thanks!
... View more