Member since
03-30-2016
11
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1865 | 10-14-2016 03:56 PM |
10-14-2016
03:56 PM
Thanks Santhosh B Gowda I solved it by deleting old disabling and enabling kerberos, generating new key tabs for the whole components and manually creating key tabs for Yarn App timeline server
... View more
10-12-2016
03:48 PM
Hi Santhosh, Here is the result of the list & kinit commands, it seems successful. but still can't start hive2server [root@10-0-0-81 ~]# klist -kt /etc/security/keytabs/hive.service.keytab Keytab name: FILE:/etc/security/keytabs/hive.service.keytab KVNO Timestamp Principal ---- ----------------- -------------------------------------------------------- 3 10/12/16 01:55:33 hive/10-0-0-80.cluster@CLUSTER.COM 3 10/12/16 01:55:33 hive/10-0-0-80.cluster@CLUSTER.COM 3 10/12/16 01:55:33 hive/10-0-0-80.cluster@CLUSTER.COM 3 10/12/16 01:55:33 hive/10-0-0-80.cluster@CLUSTER.COM 3 10/12/16 01:55:33 hive/10-0-0-80.cluster@CLUSTER.COM
... View more
10-12-2016
05:23 AM
<title>Error 403 GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos credentails)</title> </head>
<body><h2>HTTP ERROR 403</h2>
<p>Problem accessing /webhdfs/v1/ats/done. Reason:
<pre> GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos credentails)</pre></p><hr /><i><small>Powered by Jetty://</small></i><br/>
<br/>
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Hive
-
Apache YARN
09-12-2016
05:43 AM
Thanks Jonathan, i have been battling with this for a week, this tip helped, i had symlink errors with accumulo-client installation of HDP 2.4
... View more
07-15-2016
09:57 AM
Thanks Artem that worked
... View more
06-17-2016
09:44 AM
stderr:
Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 64, in <module>
MysqlServer().execute()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 218, in execute
method(env)
File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 33, in install
self.install_packages(env, exclude_packages=params.hive_exclude_packages)
File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 376, in install_packages
Package(name)
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 157, in __init__
self.env.run()
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
self.run_action(resource, action)
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
provider_action()
File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/__init__.py", line 45, in action_install
self.install_package(package_name, self.resource.use_repos, self.resource.skip_repos)
File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/yumrpm.py", line 49, in install_package
shell.checked_call(cmd, sudo=True, logoutput=self.get_logoutput())
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 70, in inner
result = function(command, **kwargs)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 92, in checked_call
tries=tries, try_sleep=try_sleep)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 140, in _call_wrapper
result = _call(command, **kwargs_copy)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 291, in _call
raise Fail(err_msg)
resource_management.core.exceptions.Fail: Execution of '/usr/bin/yum -d 0 -e 0 -y install mysql-community-server' returned 1. Removing mariadb-server.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-server.x86_64 0:5.6.31-2.el7 - u
Removing mariadb.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-client.x86_64 0:5.6.31-2.el7 - u
Removing mariadb-libs.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-libs.x86_64 0:5.6.31-2.el7 - u
Error: Package: akonadi-mysql-1.9.2-4.el7.x86_64 (@anaconda)
Requires: mariadb-server
Removing: 1:mariadb-server-5.5.44-2.el7.centos.x86_64 (@anaconda)
mariadb-server = 1:5.5.44-2.el7.centos
Obsoleted By: mysql-community-server-5.6.31-2.el7.x86_64 (mysql56-community)
Not found
Updated By: 1:mariadb-server-5.5.47-1.el7_2.x86_64 (updates)
mariadb-server = 1:5.5.47-1.el7_2
You could try using --skip-broken to work around the problem
You could try running: rpm -Va --nofiles --nodigest
stdout:
2016-06-16 19:43:02,262 - Directory['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts/'] {'recursive': True}
2016-06-16 19:43:02,263 - File['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts//jce_policy-8.zip'] {'content': DownloadSource('http://Node1.cluster.com:8080/resources//jce_policy-8.zip')}
2016-06-16 19:43:02,263 - Not downloading the file from http://Node1.cluster.com:8080/resources//jce_policy-8.zip, because /var/lib/ambari-agent/data/tmp/jce_policy-8.zip already exists
2016-06-16 19:43:02,264 - Group['spark'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['hadoop'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['users'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['knox'] {'ignore_failures': False}
2016-06-16 19:43:02,267 - User['hive'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,268 - User['storm'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,268 - User['zookeeper'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,269 - User['oozie'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,270 - User['atlas'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,271 - User['ams'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,271 - User['falcon'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,272 - User['tez'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,273 - User['accumulo'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,273 - User['mahout'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,274 - User['spark'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,275 - User['ambari-qa'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,275 - User['flume'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,276 - User['kafka'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,277 - User['hdfs'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,278 - User['sqoop'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,278 - User['yarn'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,279 - User['mapred'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,280 - User['hbase'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,280 - User['knox'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,281 - User['hcat'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,282 - File['/var/lib/ambari-agent/data/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2016-06-16 19:43:02,283 - Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2016-06-16 19:43:02,290 - Skipping Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2016-06-16 19:43:02,290 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
2016-06-16 19:43:02,291 - File['/var/lib/ambari-agent/data/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2016-06-16 19:43:02,292 - Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2016-06-16 19:43:02,297 - Skipping Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2016-06-16 19:43:02,298 - Group['hdfs'] {'ignore_failures': False}
2016-06-16 19:43:02,298 - User['hdfs'] {'ignore_failures': False, 'groups': [u'hadoop', u'hdfs']}
2016-06-16 19:43:02,299 - Directory['/etc/hadoop'] {'mode': 0755}
2016-06-16 19:43:02,311 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2016-06-16 19:43:02,325 - Repository['HDP-2.3'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP/centos7/2.x/updates/2.3.4.7', 'action': ['create'], 'components': [u'HDP', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP', 'mirror_list': None}
2016-06-16 19:43:02,333 - File['/etc/yum.repos.d/HDP.repo'] {'content': InlineTemplate(...)}
2016-06-16 19:43:02,334 - Repository['HDP-UTILS-1.1.0.20'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.20/repos/centos7', 'action': ['create'], 'components': [u'HDP-UTILS', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP-UTILS', 'mirror_list': None}
2016-06-16 19:43:02,337 - File['/etc/yum.repos.d/HDP-UTILS.repo'] {'content': InlineTemplate(...)}
2016-06-16 19:43:02,338 - Package['unzip'] {}
2016-06-16 19:43:02,443 - Skipping installation of existing package unzip
2016-06-16 19:43:02,443 - Package['curl'] {}
2016-06-16 19:43:02,461 - Skipping installation of existing package curl
2016-06-16 19:43:02,461 - Package['hdp-select'] {}
2016-06-16 19:43:02,479 - Skipping installation of existing package hdp-select
2016-06-16 19:43:02,479 - Directory['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts/'] {'recursive': True}
2016-06-16 19:43:02,480 - File['/var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'] {'content': DownloadSource('http://Node1.cluster.com:8080/resources//jdk-8u40-linux-x64.tar.gz'), 'not_if': 'test -f /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'}
2016-06-16 19:43:02,484 - Skipping File['/var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'] due to not_if
2016-06-16 19:43:02,485 - Directory['/usr/jdk64'] {}
2016-06-16 19:43:02,485 - Execute['('chmod', 'a+x', u'/usr/jdk64')'] {'not_if': 'test -e /usr/jdk64/jdk1.8.0_40/bin/java', 'sudo': True}
2016-06-16 19:43:02,490 - Skipping Execute['('chmod', 'a+x', u'/usr/jdk64')'] due to not_if
2016-06-16 19:43:02,491 - Execute['mkdir -p /var/lib/ambari-agent/data/tmp/jdk && cd /var/lib/ambari-agent/data/tmp/jdk && tar -xf /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz && ambari-sudo.sh cp -rp /var/lib/ambari-agent/data/tmp/jdk/* /usr/jdk64'] {'not_if': 'test -e /usr/jdk64/jdk1.8.0_40/bin/java'}
2016-06-16 19:43:02,495 - Skipping Execute['mkdir -p /var/lib/ambari-agent/data/tmp/jdk && cd /var/lib/ambari-agent/data/tmp/jdk && tar -xf /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz && ambari-sudo.sh cp -rp /var/lib/ambari-agent/data/tmp/jdk/* /usr/jdk64'] due to not_if
2016-06-16 19:43:02,496 - File['/usr/jdk64/jdk1.8.0_40/bin/java'] {'mode': 0755, 'cd_access': 'a'}
2016-06-16 19:43:02,496 - Execute['('chgrp', '-R', u'hadoop', u'/usr/jdk64/jdk1.8.0_40')'] {'sudo': True}
2016-06-16 19:43:02,510 - Execute['('chown', '-R', 'root', u'/usr/jdk64/jdk1.8.0_40')'] {'sudo': True}
2016-06-16 19:43:02,772 - Package['atlas-metadata*-hive-plugin'] {}
2016-06-16 19:43:02,873 - Skipping installation of existing package atlas-metadata*-hive-plugin
2016-06-16 19:43:02,874 - Package['mysql-community-release'] {}
2016-06-16 19:43:02,892 - Installing package mysql-community-release ('/usr/bin/yum -d 0 -e 0 -y install mysql-community-release')
2016-06-16 19:43:05,762 - Package['mysql-community-server'] {}
2016-06-16 19:43:05,781 - Installing package mysql-community-server ('/usr/bin/yum -d 0 -e 0 -y install mysql-community-server')
... View more
Labels: