Created 12-13-2018 03:01 PM
hi, i am trying to install hadoop via ambari but i get this error
stderr:
Traceback (most recent call last):
File "/usr/lib/ambari-agent/lib/resource_management/core/source.py", line 195, in get_content
web_file = opener.open(req)
File "/usr/lib64/python2.7/urllib2.py", line 437, in open
response = meth(req, response)
File "/usr/lib64/python2.7/urllib2.py", line 550, in http_response
'http', request, response, code, msg, hdrs)
File "/usr/lib64/python2.7/urllib2.py", line 475, in error
return self._call_chain(*args)
File "/usr/lib64/python2.7/urllib2.py", line 409, in _call_chain
result = func(*args)
File "/usr/lib64/python2.7/urllib2.py", line 558, in http_error_default
raise HTTPError(req.get_full_url(), code, msg, hdrs, fp)
HTTPError: HTTP Error 404: Not Found
The above exception was the cause of the following exception:
Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/HIVE/package/scripts/hive_client.py", line 60, in
HiveClient().execute()
File "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", line 351, in execute
method(env)
File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/HIVE/package/scripts/hive_client.py", line 40, in install
self.configure(env)
File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/HIVE/package/scripts/hive_client.py", line 48, in configure
hive(name='client')
File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/HIVE/package/scripts/hive.py", line 114, in hive
jdbc_connector(params.hive_jdbc_target, params.hive_previous_jdbc_jar)
File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/HIVE/package/scripts/hive.py", line 630, in jdbc_connector
File(params.downloaded_custom_connector, content = DownloadSource(params.driver_curl_source))
File "/usr/lib/ambari-agent/lib/resource_management/core/base.py", line 166, in __init__
self.env.run()
File "/usr/lib/ambari-agent/lib/resource_management/core/environment.py", line 160, in run
self.run_action(resource, action)
File "/usr/lib/ambari-agent/lib/resource_management/core/environment.py", line 124, in run_action
provider_action()
File "/usr/lib/ambari-agent/lib/resource_management/core/providers/system.py", line 123, in action_create
content = self._get_content()
File "/usr/lib/ambari-agent/lib/resource_management/core/providers/system.py", line 160, in _get_content
return content()
File "/usr/lib/ambari-agent/lib/resource_management/core/source.py", line 52, in __call__
return self.get_content()
File "/usr/lib/ambari-agent/lib/resource_management/core/source.py", line 197, in get_content
raise Fail("Failed to download file from {0} due to HTTP error: {1}".format(self.url, str(ex)))
resource_management.core.exceptions.Fail: Failed to download file from http://ambari.hadoop.uom.gr:8080/resources/mysql-connector-java.jar due to HTTP error: HTTP Error 404: Not Found
stdout:
2018-12-13 16:32:15,104 - Stack Feature Version Info: Cluster Stack=3.0, Command Stack=None, Command Version=None -> 3.0
2018-12-13 16:32:15,107 - Using hadoop conf dir: /usr/hdp/3.0.1.0-187/hadoop/conf
2018-12-13 16:32:15,108 - Group['livy'] {}
2018-12-13 16:32:15,109 - Group['spark'] {}
2018-12-13 16:32:15,109 - Group['hdfs'] {}
2018-12-13 16:32:15,109 - Group['zeppelin'] {}
2018-12-13 16:32:15,109 - Group['hadoop'] {}
2018-12-13 16:32:15,109 - Group['users'] {}
2018-12-13 16:32:15,109 - Group['knox'] {}
2018-12-13 16:32:15,109 - User['yarn-ats'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,110 - User['hive'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,111 - User['storm'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,111 - User['infra-solr'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,112 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,112 - User['oozie'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'users'], 'uid': None}
2018-12-13 16:32:15,113 - User['atlas'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,113 - User['ams'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,114 - User['tez'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'users'], 'uid': None}
2018-12-13 16:32:15,114 - User['zeppelin'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['zeppelin', 'hadoop'], 'uid': None}
2018-12-13 16:32:15,115 - User['accumulo'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,116 - User['livy'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['livy', 'hadoop'], 'uid': None}
2018-12-13 16:32:15,116 - User['druid'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,117 - User['spark'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['spark', 'hadoop'], 'uid': None}
2018-12-13 16:32:15,117 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'users'], 'uid': None}
2018-12-13 16:32:15,118 - User['kafka'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,118 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop'], 'uid': None}
2018-12-13 16:32:15,119 - User['sqoop'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,119 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,120 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,121 - User['hbase'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2018-12-13 16:32:15,121 - User['knox'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'knox'], 'uid': None}
2018-12-13 16:32:15,122 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2018-12-13 16:32:15,123 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2018-12-13 16:32:15,126 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] due to not_if
2018-12-13 16:32:15,126 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'}
2018-12-13 16:32:15,127 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2018-12-13 16:32:15,128 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2018-12-13 16:32:15,129 - call['/var/lib/ambari-agent/tmp/changeUid.sh hbase'] {}
2018-12-13 16:32:15,133 - call returned (0, '1020')
2018-12-13 16:32:15,134 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1020'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2018-12-13 16:32:15,137 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1020'] due to not_if
2018-12-13 16:32:15,137 - Group['hdfs'] {}
2018-12-13 16:32:15,137 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop', u'hdfs']}
2018-12-13 16:32:15,138 - FS Type: HDFS
2018-12-13 16:32:15,138 - Directory['/etc/hadoop'] {'mode': 0755}
2018-12-13 16:32:15,147 - File['/usr/hdp/3.0.1.0-187/hadoop/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2018-12-13 16:32:15,147 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777}
2018-12-13 16:32:15,159 - Repository['HDP-3.0-repo-1'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP/centos7/3.x/updates/3.0.1.0', 'action': ['prepare'], 'components': [u'HDP', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-hdp-1', 'mirror_list': None}
2018-12-13 16:32:15,163 - Repository['HDP-3.0-GPL-repo-1'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP-GPL/centos7/3.x/updates/3.0.1.0', 'action': ['prepare'], 'components': [u'HDP-GPL', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-hdp-1', 'mirror_list': None}
2018-12-13 16:32:15,165 - Repository['HDP-UTILS-1.1.0.22-repo-1'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.22/repos/centos7', 'action': ['prepare'], 'components': [u'HDP-UTILS', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-hdp-1', 'mirror_list': None}
2018-12-13 16:32:15,166 - Repository[None] {'action': ['create']}
2018-12-13 16:32:15,167 - File['/tmp/tmpbESTEN'] {'content': '[HDP-3.0-repo-1]\nname=HDP-3.0-repo-1\nbaseurl=http://public-repo-1.hortonworks.com/HDP/centos7/3.x/updates/3.0.1.0\n\npath=/\nenabled=1\ngpgcheck=0\n[HDP-3.0-GPL-repo-1]\nname=HDP-3.0-GPL-repo-1\nbaseurl=http://public-repo-1.hortonworks.com/HDP-GPL/centos7/3.x/updates/3.0.1.0\n\npath=/\nenabled=1\ngpgcheck=0\n[HDP-UTILS-1.1.0.22-repo-1]\nname=HDP-UTILS-1.1.0.22-repo-1\nbaseurl=http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.22/repos/centos7\n\npath=/\nenabled=1\ngpgcheck=0'}
2018-12-13 16:32:15,167 - Writing File['/tmp/tmpbESTEN'] because contents don't match
2018-12-13 16:32:15,167 - File['/tmp/tmptU9nw5'] {'content': StaticFile('/etc/yum.repos.d/ambari-hdp-1.repo')}
2018-12-13 16:32:15,168 - Writing File['/tmp/tmptU9nw5'] because contents don't match
2018-12-13 16:32:15,168 - Package['unzip'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2018-12-13 16:32:15,221 - Skipping installation of existing package unzip
2018-12-13 16:32:15,221 - Package['curl'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2018-12-13 16:32:15,227 - Skipping installation of existing package curl
2018-12-13 16:32:15,227 - Package['hdp-select'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2018-12-13 16:32:15,233 - Skipping installation of existing package hdp-select
2018-12-13 16:32:15,236 - The repository with version 3.0.1.0-187 for this command has been marked as resolved. It will be used to report the version of the component which was installed
2018-12-13 16:32:15,424 - Using hadoop conf dir: /usr/hdp/3.0.1.0-187/hadoop/conf
2018-12-13 16:32:15,430 - call['ambari-python-wrap /usr/bin/hdp-select status hive-server2'] {'timeout': 20}
2018-12-13 16:32:15,445 - call returned (0, 'hive-server2 - 3.0.1.0-187')
2018-12-13 16:32:15,447 - Stack Feature Version Info: Cluster Stack=3.0, Command Stack=None, Command Version=None -> 3.0
2018-12-13 16:32:15,461 - File['/var/lib/ambari-agent/cred/lib/CredentialUtil.jar'] {'content': DownloadSource('http://ambari.hadoop.uom.gr:8080/resources/CredentialUtil.jar'), 'mode': 0755}
2018-12-13 16:32:15,462 - Not downloading the file from http://ambari.hadoop.uom.gr:8080/resources/CredentialUtil.jar, because /var/lib/ambari-agent/tmp/CredentialUtil.jar already exists
2018-12-13 16:32:15,996 - Package['hive_3_0_1_0_187'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2018-12-13 16:32:16,051 - Skipping installation of existing package hive_3_0_1_0_187
2018-12-13 16:32:16,052 - Package['hive_3_0_1_0_187-hcatalog'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2018-12-13 16:32:16,057 - Skipping installation of existing package hive_3_0_1_0_187-hcatalog
2018-12-13 16:32:16,058 - Directories to fill with configs: [u'/usr/hdp/current/hive-client/conf']
2018-12-13 16:32:16,059 - Directory['/etc/hive/3.0.1.0-187/0'] {'owner': 'hive', 'group': 'hadoop', 'create_parents': True, 'mode': 0755}
2018-12-13 16:32:16,059 - XmlConfig['mapred-site.xml'] {'group': 'hadoop', 'conf_dir': '/etc/hive/3.0.1.0-187/0', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hive', 'configurations': ...}
2018-12-13 16:32:16,066 - Generating config: /etc/hive/3.0.1.0-187/0/mapred-site.xml
2018-12-13 16:32:16,066 - File['/etc/hive/3.0.1.0-187/0/mapred-site.xml'] {'owner': 'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
2018-12-13 16:32:16,092 - File['/etc/hive/3.0.1.0-187/0/hive-default.xml.template'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,093 - File['/etc/hive/3.0.1.0-187/0/hive-env.sh.template'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0755}
2018-12-13 16:32:16,095 - File['/etc/hive/3.0.1.0-187/0/llap-daemon-log4j2.properties'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,096 - File['/etc/hive/3.0.1.0-187/0/llap-cli-log4j2.properties'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,098 - File['/etc/hive/3.0.1.0-187/0/hive-log4j2.properties'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,099 - File['/etc/hive/3.0.1.0-187/0/hive-exec-log4j2.properties'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,100 - File['/etc/hive/3.0.1.0-187/0/beeline-log4j2.properties'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,100 - XmlConfig['beeline-site.xml'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hive/3.0.1.0-187/0', 'configurations': {'beeline.hs2.jdbc.url.container': u'jdbc:hive2://namenode.hadoop.uom.gr:2181,resourcemanager.hadoop.uom.gr:2181,hbasemaster.hadoop.uom.gr:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2', 'beeline.hs2.jdbc.url.default': 'container'}}
2018-12-13 16:32:16,106 - Generating config: /etc/hive/3.0.1.0-187/0/beeline-site.xml
2018-12-13 16:32:16,106 - File['/etc/hive/3.0.1.0-187/0/beeline-site.xml'] {'owner': 'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
2018-12-13 16:32:16,108 - File['/etc/hive/3.0.1.0-187/0/parquet-logging.properties'] {'content': ..., 'owner': 'hive', 'group': 'hadoop', 'mode': 0644}
2018-12-13 16:32:16,108 - File['/usr/hdp/current/hive-client/conf/hive-site.jceks'] {'content': StaticFile('/var/lib/ambari-agent/cred/conf/hive_client/hive-site.jceks'), 'owner': 'hive', 'group': 'hadoop', 'mode': 0640}
2018-12-13 16:32:16,108 - Writing File['/usr/hdp/current/hive-client/conf/hive-site.jceks'] because contents don't match
2018-12-13 16:32:16,109 - XmlConfig['hive-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hive-client/conf', 'mode': 0644, 'configuration_attributes': {u'hidden': {u'javax.jdo.option.ConnectionPassword': u'HIVE_CLIENT,CONFIG_DOWNLOAD'}}, 'owner': 'hive', 'configurations': ...}
2018-12-13 16:32:16,113 - Generating config: /usr/hdp/current/hive-client/conf/hive-site.xml
2018-12-13 16:32:16,114 - File['/usr/hdp/current/hive-client/conf/hive-site.xml'] {'owner': 'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
2018-12-13 16:32:16,211 - Generating Atlas Hook config file /usr/hdp/current/hive-client/conf/atlas-application.properties
2018-12-13 16:32:16,212 - PropertiesFile['/usr/hdp/current/hive-client/conf/atlas-application.properties'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0644, 'properties': ...}
2018-12-13 16:32:16,214 - Generating properties file: /usr/hdp/current/hive-client/conf/atlas-application.properties
2018-12-13 16:32:16,214 - File['/usr/hdp/current/hive-client/conf/atlas-application.properties'] {'owner': 'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
2018-12-13 16:32:16,223 - Writing File['/usr/hdp/current/hive-client/conf/atlas-application.properties'] because contents don't match
2018-12-13 16:32:16,226 - File['/usr/hdp/current/hive-client/conf/hive-env.sh'] {'content': InlineTemplate(...), 'owner': 'hive', 'group': 'hadoop', 'mode': 0755}
2018-12-13 16:32:16,227 - Directory['/etc/security/limits.d'] {'owner': 'root', 'create_parents': True, 'group': 'root'}
2018-12-13 16:32:16,230 - File['/etc/security/limits.d/hive.conf'] {'content': Template('hive.conf.j2'), 'owner': 'root', 'group': 'root', 'mode': 0644}
2018-12-13 16:32:16,231 - File['/usr/lib/ambari-agent/DBConnectionVerification.jar'] {'content': DownloadSource('http://ambari.hadoop.uom.gr:8080/resources/DBConnectionVerification.jar'), 'mode': 0644}
2018-12-13 16:32:16,231 - Not downloading the file from http://ambari.hadoop.uom.gr:8080/resources/DBConnectionVerification.jar, because /var/lib/ambari-agent/tmp/DBConnectionVerification.jar already exists
2018-12-13 16:32:16,231 - File['/var/lib/ambari-agent/tmp/mysql-connector-java.jar'] {'content': DownloadSource('http://ambari.hadoop.uom.gr:8080/resources/mysql-connector-java.jar')}
2018-12-13 16:32:16,231 - Downloading the file from http://ambari.hadoop.uom.gr:8080/resources/mysql-connector-java.jar
2018-12-13 16:32:16,241 - The repository with version 3.0.1.0-187 for this command has been marked as resolved. It will be used to report the version of the component which was installed
Command failed after 1 tries
Created 12-13-2018 04:18 PM
Hi @Bill Ferris,
For the logs i see you are having this exception :
resource_management.core.exceptions.Fail:Failed to download file from http://ambari.hadoop.uom.gr:8080/resources/mysql-connector-java.jar due to HTTP error: HTTP Error 404: Not Found
Can you please try the following in ambari-server and see if this helps ?
yum install mysql-connection-java -y
(OR) if you are downloading the mysql-connector-java JAR from some tar.gz archive then please make sure to check the following locations and create the symlinks something like following to point to your jar.
.Then you should find some symlink as following:
Example:
# ls -l /usr/share/java/mysql-connector-java.jarl rwxrwxrwx 1 root root 31 Apr 19 2017 /usr/share/java/mysql-connector-java.jar -> mysql-connector-java-5.1.17.jar
So now ambari knows how to find this jar. The JAR can be found hereafter
# ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar # ls -l /var/lib/ambari-server/resources/mysql-connector-java.jar -rw-r--r-- 1 root root 819803 Sep 28 19:52 /var/lib/ambari-server/resources/mysql-connector-java.jar
Please accept this answer if you found this helpful
Created 12-13-2018 04:18 PM
Hi @Bill Ferris,
For the logs i see you are having this exception :
resource_management.core.exceptions.Fail:Failed to download file from http://ambari.hadoop.uom.gr:8080/resources/mysql-connector-java.jar due to HTTP error: HTTP Error 404: Not Found
Can you please try the following in ambari-server and see if this helps ?
yum install mysql-connection-java -y
(OR) if you are downloading the mysql-connector-java JAR from some tar.gz archive then please make sure to check the following locations and create the symlinks something like following to point to your jar.
.Then you should find some symlink as following:
Example:
# ls -l /usr/share/java/mysql-connector-java.jarl rwxrwxrwx 1 root root 31 Apr 19 2017 /usr/share/java/mysql-connector-java.jar -> mysql-connector-java-5.1.17.jar
So now ambari knows how to find this jar. The JAR can be found hereafter
# ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar # ls -l /var/lib/ambari-server/resources/mysql-connector-java.jar -rw-r--r-- 1 root root 819803 Sep 28 19:52 /var/lib/ambari-server/resources/mysql-connector-java.jar
Please accept this answer if you found this helpful
Created 12-13-2018 05:33 PM
thanks for the reply, i have centos 7.6 and get this output
no package mysql-connection-java available
nothing to do
Created 12-13-2018 05:42 PM
Hi @Bill Ferris,
Please download mysql connector jar from here if you cannot install it via yum command : https://dev.mysql.com/downloads/connector/j/
Please login and accept answer as helpfull if it worked for you