Support Questions

Find answers, ask questions, and share your expertise
Announcements
Celebrating as our community reaches 100,000 members! Thank you!

Hello, i am installing a singe node cluster, and now installing ambari agent manually, when it gets to stage 9, it installs all other services but stops during installation of mysql server, then installation stops and gives error, pls help me.

avatar
stderr: 
Traceback (most recent call last):
  File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 64, in <module>
  MysqlServer().execute()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 218, in execute
  method(env)
  File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 33, in install
  self.install_packages(env, exclude_packages=params.hive_exclude_packages)
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 376, in install_packages
  Package(name)
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 157, in __init__
  self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
  self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
  provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/__init__.py", line 45, in action_install
  self.install_package(package_name, self.resource.use_repos, self.resource.skip_repos)
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/yumrpm.py", line 49, in install_package
  shell.checked_call(cmd, sudo=True, logoutput=self.get_logoutput())
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 70, in inner
  result = function(command, **kwargs)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 92, in checked_call
  tries=tries, try_sleep=try_sleep)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 140, in _call_wrapper
  result = _call(command, **kwargs_copy)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 291, in _call
  raise Fail(err_msg)
resource_management.core.exceptions.Fail: Execution of '/usr/bin/yum -d 0 -e 0 -y install mysql-community-server' returned 1. Removing mariadb-server.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-server.x86_64 0:5.6.31-2.el7 - u
Removing mariadb.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-client.x86_64 0:5.6.31-2.el7 - u
Removing mariadb-libs.x86_64 1:5.5.47-1.el7_2 - u due to obsoletes from mysql-community-libs.x86_64 0:5.6.31-2.el7 - u
Error: Package: akonadi-mysql-1.9.2-4.el7.x86_64 (@anaconda)
  Requires: mariadb-server
  Removing: 1:mariadb-server-5.5.44-2.el7.centos.x86_64 (@anaconda)
  mariadb-server = 1:5.5.44-2.el7.centos
  Obsoleted By: mysql-community-server-5.6.31-2.el7.x86_64 (mysql56-community)
  Not found
  Updated By: 1:mariadb-server-5.5.47-1.el7_2.x86_64 (updates)
  mariadb-server = 1:5.5.47-1.el7_2
 You could try using --skip-broken to work around the problem
 You could try running: rpm -Va --nofiles --nodigest
 stdout:
2016-06-16 19:43:02,262 - Directory['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts/'] {'recursive': True}
2016-06-16 19:43:02,263 - File['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts//jce_policy-8.zip'] {'content': DownloadSource('http://Node1.cluster.com:8080/resources//jce_policy-8.zip')}
2016-06-16 19:43:02,263 - Not downloading the file from http://Node1.cluster.com:8080/resources//jce_policy-8.zip, because /var/lib/ambari-agent/data/tmp/jce_policy-8.zip already exists
2016-06-16 19:43:02,264 - Group['spark'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['hadoop'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['users'] {'ignore_failures': False}
2016-06-16 19:43:02,266 - Group['knox'] {'ignore_failures': False}
2016-06-16 19:43:02,267 - User['hive'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,268 - User['storm'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,268 - User['zookeeper'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,269 - User['oozie'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,270 - User['atlas'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,271 - User['ams'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,271 - User['falcon'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,272 - User['tez'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,273 - User['accumulo'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,273 - User['mahout'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,274 - User['spark'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,275 - User['ambari-qa'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
2016-06-16 19:43:02,275 - User['flume'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,276 - User['kafka'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,277 - User['hdfs'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,278 - User['sqoop'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,278 - User['yarn'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,279 - User['mapred'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,280 - User['hbase'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,280 - User['knox'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,281 - User['hcat'] {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
2016-06-16 19:43:02,282 - File['/var/lib/ambari-agent/data/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2016-06-16 19:43:02,283 - Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2016-06-16 19:43:02,290 - Skipping Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2016-06-16 19:43:02,290 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
2016-06-16 19:43:02,291 - File['/var/lib/ambari-agent/data/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2016-06-16 19:43:02,292 - Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2016-06-16 19:43:02,297 - Skipping Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2016-06-16 19:43:02,298 - Group['hdfs'] {'ignore_failures': False}
2016-06-16 19:43:02,298 - User['hdfs'] {'ignore_failures': False, 'groups': [u'hadoop', u'hdfs']}
2016-06-16 19:43:02,299 - Directory['/etc/hadoop'] {'mode': 0755}
2016-06-16 19:43:02,311 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2016-06-16 19:43:02,325 - Repository['HDP-2.3'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP/centos7/2.x/updates/2.3.4.7', 'action': ['create'], 'components': [u'HDP', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP', 'mirror_list': None}
2016-06-16 19:43:02,333 - File['/etc/yum.repos.d/HDP.repo'] {'content': InlineTemplate(...)}
2016-06-16 19:43:02,334 - Repository['HDP-UTILS-1.1.0.20'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.20/repos/centos7', 'action': ['create'], 'components': [u'HDP-UTILS', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP-UTILS', 'mirror_list': None}
2016-06-16 19:43:02,337 - File['/etc/yum.repos.d/HDP-UTILS.repo'] {'content': InlineTemplate(...)}
2016-06-16 19:43:02,338 - Package['unzip'] {}
2016-06-16 19:43:02,443 - Skipping installation of existing package unzip
2016-06-16 19:43:02,443 - Package['curl'] {}
2016-06-16 19:43:02,461 - Skipping installation of existing package curl
2016-06-16 19:43:02,461 - Package['hdp-select'] {}
2016-06-16 19:43:02,479 - Skipping installation of existing package hdp-select
2016-06-16 19:43:02,479 - Directory['/var/lib/ambari-agent/data/tmp/AMBARI-artifacts/'] {'recursive': True}
2016-06-16 19:43:02,480 - File['/var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'] {'content': DownloadSource('http://Node1.cluster.com:8080/resources//jdk-8u40-linux-x64.tar.gz'), 'not_if': 'test -f /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'}
2016-06-16 19:43:02,484 - Skipping File['/var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz'] due to not_if
2016-06-16 19:43:02,485 - Directory['/usr/jdk64'] {}
2016-06-16 19:43:02,485 - Execute['('chmod', 'a+x', u'/usr/jdk64')'] {'not_if': 'test -e /usr/jdk64/jdk1.8.0_40/bin/java', 'sudo': True}
2016-06-16 19:43:02,490 - Skipping Execute['('chmod', 'a+x', u'/usr/jdk64')'] due to not_if
2016-06-16 19:43:02,491 - Execute['mkdir -p /var/lib/ambari-agent/data/tmp/jdk && cd /var/lib/ambari-agent/data/tmp/jdk && tar -xf /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz && ambari-sudo.sh cp -rp /var/lib/ambari-agent/data/tmp/jdk/* /usr/jdk64'] {'not_if': 'test -e /usr/jdk64/jdk1.8.0_40/bin/java'}
2016-06-16 19:43:02,495 - Skipping Execute['mkdir -p /var/lib/ambari-agent/data/tmp/jdk && cd /var/lib/ambari-agent/data/tmp/jdk && tar -xf /var/lib/ambari-agent/data/tmp/jdk-8u40-linux-x64.tar.gz && ambari-sudo.sh cp -rp /var/lib/ambari-agent/data/tmp/jdk/* /usr/jdk64'] due to not_if
2016-06-16 19:43:02,496 - File['/usr/jdk64/jdk1.8.0_40/bin/java'] {'mode': 0755, 'cd_access': 'a'}
2016-06-16 19:43:02,496 - Execute['('chgrp', '-R', u'hadoop', u'/usr/jdk64/jdk1.8.0_40')'] {'sudo': True}
2016-06-16 19:43:02,510 - Execute['('chown', '-R', 'root', u'/usr/jdk64/jdk1.8.0_40')'] {'sudo': True}
2016-06-16 19:43:02,772 - Package['atlas-metadata*-hive-plugin'] {}
2016-06-16 19:43:02,873 - Skipping installation of existing package atlas-metadata*-hive-plugin
2016-06-16 19:43:02,874 - Package['mysql-community-release'] {}
2016-06-16 19:43:02,892 - Installing package mysql-community-release ('/usr/bin/yum -d 0 -e 0 -y install mysql-community-release')
2016-06-16 19:43:05,762 - Package['mysql-community-server'] {}
2016-06-16 19:43:05,781 - Installing package mysql-community-server ('/usr/bin/yum -d 0 -e 0 -y install mysql-community-server')
1 ACCEPTED SOLUTION

avatar
Master Mentor

Run MySQL install on that machine manually to resolve dependency issues, you have conflicts between MariaDB and MySQL

View solution in original post

3 REPLIES 3

avatar
Master Mentor

Run MySQL install on that machine manually to resolve dependency issues, you have conflicts between MariaDB and MySQL

avatar

Thanks Artem that worked

avatar
Master Mentor

Glad that worked. Please accept the answer.