Member since
04-08-2018
64
Posts
2
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7343 | 05-04-2018 05:01 PM | |
6656 | 04-16-2018 10:13 AM |
05-07-2018
01:45 PM
I have just tested it. It worked fine! Thank you!
... View more
05-07-2018
01:30 PM
Yes, sure. Sorry, I was actually referring to "hdfs://eureambarimaster1.local.eurecat.org:8020/user/hdfs/test/df.parquet" Let me test it.
... View more
05-07-2018
01:05 PM
I think that this is the reason. If I login as HDFS user and run "hdfs dfs -chown -R centos /home/centos/test", then it says that this directory does not exist. I created this directory as HDFS user and then changed permissions to centos. Should I write a parquet file to the full path?: df.coalesce(1).write.format("parquet").save("hdfs://eureambarimaster1.local.eurecat.org:8020/user/hdfs/test")
... View more
05-07-2018
12:43 PM
Maybe the problem is that I run Spark program in Yarn cluster mode? It means that the driver can be running in any of the machines of the cluster. So, probably I should run "chown -R centos:centos ..." in each machine or do ".coalesce(1)"?
... View more
05-07-2018
12:37 PM
The output of "id": uid=1000(centos) gid=1000(centos) groups=1000(centos),4(adm),10(wheel),190(systemd-journal) I executed "chown -R centos:centos /home/centos/test" but still get the same error: 18/05/07 12:06:28 ERROR ApplicationMaster: User class threw exception: org.apache.hadoop.security.AccessControlException: Permission denied: user=centos, access=WRITE, inode="/home/centos/test/df.parquet/_temporary/0":hdfs:hdfs:drwxr-xr-x This is the output of "ls -la" executed in "/home/centos": total 36236
drwx------. 4 centos centos 4096 May 7 12:34 .
drwxr-xr-x. 15 root root 4096 Apr 16 18:41 ..
-rw-------. 1 centos centos 13781 May 7 11:26 .bash_history
-rw-r--r--. 1 centos centos 18 Mar 5 2015 .bash_logout
-rw-r--r--. 1 centos centos 193 Mar 5 2015 .bash_profile
-rw-r--r--. 1 centos centos 231 Mar 5 2015 .bashrc
-rw-rw-r-- 1 centos centos 47 May 7 11:38 .scala_history
drwx------. 2 centos centos 46 May 2 07:57 .ssh
drwxrwxr-x 4 centos centos 144 May 7 11:42 test
... View more
05-07-2018
09:38 AM
I want to save DataFrame on disk: df.write.format("parquet").save("/home/centos/test/df.parquet") I get the following error, which says that the user "centos" does not have write permissions: 18/05/07 09:18:08 ERROR ApplicationMaster: User class threw exception: org.apache.hadoop.security.AccessControlException: Permission denied: user=centos, access=WRITE, inode="/home/centos/test/df.parquet/_temporary/0":hdfs:hdfs:drwxr-xr-x This is how I run spark-submit command: spark-submit --master yarn --deploy-mode cluster --driver-memory 6g --executor-cores 2 --num-executors 2 --executor-memory 4g --class org.test.MyProcessor mytest.jar
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Spark
05-07-2018
08:45 AM
I re-submitted Spark job and now it works fine. The problem was that I submitted Spark job before changing permissions.
... View more
05-07-2018
08:34 AM
Yes, sure. Please see attached more screenshots from the RM UI. Thanks.
... View more
05-07-2018
06:42 AM
However, the application with such Id exists in ResourceManager. Please see the attached screenshot.
... View more
05-07-2018
06:39 AM
Thank you. I did exactly what you suggested, but I still get the same error. The directory "app_logs/centos" has ownership: centos hdfs: 18/05/07 06:36:36 INFO client.AHSProxy: Connecting to Application History server at eureambarislave1.local.eurecat.org/192.168.0.10:10200
File /app-logs/centos/logs-ifile/application_1525529485402_0020 does not exist.
File /app-logs/centos/logs/application_1525529485402_0020 does not exist.
Can not find any log file matching the pattern: [ALL] for the application: application_1525529485402_0020
Can not find the logs for the application: application_1525529485402_0020 with the appOwner: centos
... View more
05-06-2018
08:19 PM
After executing spark-submit several times, I started getting cluster capacity usage alerts. Please see the attached screenshot. I assume that the reason lies in logs. How can I clean the logs and free up the disk space?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache YARN
05-06-2018
08:11 PM
By error I deleted the folder app-logs in HDFS. After executing spark-submit, I cannot access yarn logs: [centos@eureambarimaster1]$ yarn logs -applicationId application_1525529485402_0019 -> logs1 18/05/06 20:08:14 INFO client.RMProxy: Connecting to ResourceManager at eureambarislave2.local.eurecat.org/192.168.0.15:8050 18/05/06 20:08:14 INFO client.AHSProxy: Connecting to Application History server at eureambarislave1.local.eurecat.org/192.168.0.10:10200
File /app-logs/centos/logs-ifile/application_1525529485402_0019 does not exist. How can I re-create this folder so that the user "centos" can access it?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache YARN
05-05-2018
02:27 PM
I installed Hadoop cluster with Ambari. it contains 1 master node and 2 slaves. DataNode and NodeManager are installed on each instance. Totally, I have 3 DataNodes and 3 NodeManagers. In Ambari UI I have noticed that all 3 DataNodes are alive, while only 1 NodeManager is alive (though all 3 are started). Please see the attached screenshot. I tried to add this setting to Custom hdfs-site and restarted everything, but still get the same issue: dfs.namenode.rpc-bind-host = 0.0.0.0
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
05-04-2018
05:01 PM
I had to run this command to adjust permissions: sudo -u hdfs hdfs dfs -chown centos:centos /user After this I was able to run: spark-submit --master yarn --deploy-mode cluster --driver-memory 6g--executor-memory 4g--executor-cores 2--class org.tests.GraphProcessor /path/to/graph.jar
... View more
05-04-2018
04:56 PM
In particular, if I do this way, I get the following error: Exception in thread "main" org.apache.hadoop.security.AccessControlException: Permission denied: user=centos, access=WRITE, inode="/user/centos/.sparkStaging/application_1523903913760_0007":hdfs:hdfs:drwxr-xr-x
... View more
05-04-2018
02:43 PM
If I do this way (with "sudo -u hdfs"), the jar file is invisible for hdfs user (I get an error message)). But if I run without "sudo -u hdfs", then yarn mode cannot be entered. I think that it's the matter of permissions. But it's not clear to me how to solve this issue in a most correct way. Thanks.
... View more
05-04-2018
11:36 AM
I should execute `spark-submit` in the Hadoop cluster created with Ambari. There are 3 instances: 1 master node and 2 executer nodes.
So, I logged in the master node as `centos` user and executed this command:
sudo -u hdfs spark-submit --master yarn --deploy-mode cluster --driver-memory 6g --executor-memory 4g --executor-cores 2 --class org.tests.GraphProcessor graph.jar
But I got the error message that the file graph.jar does not exist. Therefore I tried to copy this file to HDFS as follows: hdfs dfs -put graph.jar /home/hdfs/tmp However, the error is: No such file or directory: `hdfs://eureambarimaster1.local.eurecat.org:8020/home/hdfs/tmp'
... View more
Labels:
05-04-2018
11:34 AM
I should execute `spark-submit` in the Hadoop cluster created with Ambari. There are 3 instances: 1 master node and 2 executer nodes.
So, I logged in the master node as `centos` user and executed this command:
sudo -u hdfs spark-submit --master yarn --deploy-mode cluster --driver-memory 6g --executor-memory 4g --executor-cores 2 --class org.tests.GraphProcessor graph.jar
But I got the error message that the file graph.jar does not exist. Therefore I tried to copy this file to HDFS as follows: hdfs dfs -put graph.jar /home/hdfs/tmp However, the error is: No such file or directory: `hdfs://eureambarimaster1.local.eurecat.org:8020/home/hdfs/tmp'
... View more
Labels:
04-16-2018
08:54 PM
Indeed I just changed "eureambarislave1.local.eurecat.org" with "localhost" during the Hive installation process. Hive was installed successfully without any alert. I assume that "localhost" also works fine.
... View more
04-16-2018
02:02 PM
What I did manually is the following:
sudo su - wget http://repo.mysql.com/mysql-community-release-el7-5.noarch.rpm rpm -ivh mysql-community-release-el7-5.noarch.rpm Then I added Hive service using Ambari UI. According to the UI, MySQL Server was successfully installed as one of the steps of Hive installation process. If I would not do the above-mentioned steps, the installation of MySQL server would fail with the message "resource_management.core.exceptions.ExecutionFailed:Execution of '/usr/bin/yum -d 0 -e 0 -y install mysql-community-release' returned 1.Error:Nothing to do". The file "mysql-connector-java.jar" is downloaded fromhttps://dev.mysql.com/downloads/connector/j/5.1.html Then I added it to "/usr/share/java" and executed: # ls -al /usr/share/java/mysql-connector-java.jar # chmod 644 /usr/share/java/mysql-connector-java.jar # ls -l /var/lib/ambari-server/resources/mysql-connector-java.jar # ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar I cannot login to MySQL. I tried to generate password "grep 'A temporary password is generated for root@localhost' /var/log/mysqld.log |tail -1" and then do "/usr/bin/mysql_secure_installation", but I get: ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/lib/mysql/mysql.sock' (2)
... View more
04-16-2018
01:02 PM
I installed Hive with MySQL Server. The MySQL Server was installed successfully. However, I cannot start this service because of the following errors: stderr: /var/lib/ambari-agent/data/errors-148.txt Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 64, in <module>
MysqlServer().execute()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 375, in execute
method(env)
File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_server.py", line 49, in start
mysql_service(daemon_name=params.daemon_name, action='start')
File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/mysql_service.py", line 45, in mysql_service
sudo = True,
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 166, in __init__
self.env.run()
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 160, in run
self.run_action(resource, action)
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 124, in run_action
provider_action()
File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 262, in action_run
tries=self.resource.tries, try_sleep=self.resource.try_sleep)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 72, in inner
result = function(command, **kwargs)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 102, in checked_call
tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 150, in _call_wrapper
result = _call(command, **kwargs_copy)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 303, in _call
raise ExecutionFailed(err_msg, code, out, err)
resource_management.core.exceptions.ExecutionFailed: Execution of 'service mysqld start' returned 1. Redirecting to /bin/systemctl start mysqld.service
Job for mysqld.service failed because a timeout was exceeded. See "systemctl status mysqld.service" and "journalctl -xe" for details. stdout: /var/lib/ambari-agent/data/output-148.txt 2018-04-16 12:39:28,697 - Stack Feature Version Info: Cluster Stack=2.6, Command Stack=None, Command Version=2.6.4.0-91 -> 2.6.4.0-91
2018-04-16 12:39:28,714 - Using hadoop conf dir: /usr/hdp/2.6.4.0-91/hadoop/conf
2018-04-16 12:39:28,878 - Stack Feature Version Info: Cluster Stack=2.6, Command Stack=None, Command Version=2.6.4.0-91 -> 2.6.4.0-91
2018-04-16 12:39:28,884 - Using hadoop conf dir: /usr/hdp/2.6.4.0-91/hadoop/conf
2018-04-16 12:39:28,885 - Group['hdfs'] {}
2018-04-16 12:39:28,886 - Group['hadoop'] {}
2018-04-16 12:39:28,886 - Group['users'] {}
2018-04-16 12:39:28,887 - User['hive'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,888 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,888 - User['ams'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,889 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'users'], 'uid': None}
2018-04-16 12:39:28,890 - User['tez'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'users'], 'uid': None}
2018-04-16 12:39:28,891 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hdfs'], 'uid': None}
2018-04-16 12:39:28,892 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,892 - User['hcat'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,893 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': [u'hadoop'], 'uid': None}
2018-04-16 12:39:28,894 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2018-04-16 12:39:28,895 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2018-04-16 12:39:28,901 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] due to not_if
2018-04-16 12:39:28,902 - Group['hdfs'] {}
2018-04-16 12:39:28,902 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hdfs', u'hdfs']}
2018-04-16 12:39:28,902 - FS Type:
2018-04-16 12:39:28,903 - Directory['/etc/hadoop'] {'mode': 0755}
2018-04-16 12:39:28,918 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2018-04-16 12:39:28,919 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777}
2018-04-16 12:39:28,935 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'}
2018-04-16 12:39:28,942 - Skipping Execute[('setenforce', '0')] due to not_if
2018-04-16 12:39:28,942 - Directory['/var/log/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'}
2018-04-16 12:39:28,945 - Directory['/var/run/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'root', 'cd_access': 'a'}
2018-04-16 12:39:28,945 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'create_parents': True, 'cd_access': 'a'}
2018-04-16 12:39:28,949 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
2018-04-16 12:39:28,951 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'}
2018-04-16 12:39:28,957 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/log4j.properties'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
2018-04-16 12:39:28,967 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/hadoop-metrics2.properties'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2018-04-16 12:39:28,968 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
2018-04-16 12:39:28,969 - File['/usr/hdp/2.6.4.0-91/hadoop/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'}
2018-04-16 12:39:28,973 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop', 'mode': 0644}
2018-04-16 12:39:28,978 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755}
2018-04-16 12:39:28,991 - Skipping stack-select on MYSQL_SERVER because it does not exist in the stack-select package structure.
2018-04-16 12:39:29,176 - MariaDB RedHat Support: false
2018-04-16 12:39:29,181 - Using hadoop conf dir: /usr/hdp/2.6.4.0-91/hadoop/conf
2018-04-16 12:39:29,195 - call['ambari-python-wrap /usr/bin/hdp-select status hive-server2'] {'timeout': 20}
2018-04-16 12:39:29,219 - call returned (0, 'hive-server2 - 2.6.4.0-91')
2018-04-16 12:39:29,220 - Stack Feature Version Info: Cluster Stack=2.6, Command Stack=None, Command Version=2.6.4.0-91 -> 2.6.4.0-91
2018-04-16 12:39:29,255 - File['/var/lib/ambari-agent/cred/lib/CredentialUtil.jar'] {'content': DownloadSource('http://eureambarimaster1.local.eurecat.org:8080/resources/CredentialUtil.jar'), 'mode': 0755}
2018-04-16 12:39:29,257 - Not downloading the file from http://eureambarimaster1.local.eurecat.org:8080/resources/CredentialUtil.jar, because /var/lib/ambari-agent/tmp/CredentialUtil.jar already exists
2018-04-16 12:39:29,257 - checked_call[('/usr/lib/jvm/java-1.8.0-openjdk/bin/java', '-cp', u'/var/lib/ambari-agent/cred/lib/*', 'org.apache.ambari.server.credentialapi.CredentialUtil', 'get', 'javax.jdo.option.ConnectionPassword', '-provider', u'jceks://file/var/lib/ambari-agent/cred/conf/mysql_server/hive-site.jceks')] {}
2018-04-16 12:39:29,939 - checked_call returned (0, 'SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".\nSLF4J: Defaulting to no-operation (NOP) logger implementation\nSLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.\nApr 16, 2018 12:39:29 PM org.apache.hadoop.util.NativeCodeLoader <clinit>\nWARNING: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable\nEurecat@123')
2018-04-16 12:39:29,948 - Execute[('service', 'mysqld', 'start')] {'logoutput': True, 'not_if': "pgrep -l '^mysqld", 'sudo': True}
Redirecting to /bin/systemctl start mysqld.service
Job for mysqld.service failed because a timeout was exceeded. See "systemctl status mysqld.service" and "journalctl -xe" for details.
2018-04-16 12:49:30,339 - Skipping stack-select on MYSQL_SERVER because it does not exist in the stack-select package structure.
Command failed after 1 tries The logs say: 180416 12:28:20 mysqld_safe Logging to '/var/log/mysqld.log'.
180416 12:28:20 mysqld_safe Starting mysqld daemon with databases from /var/lib/mysql
2018-04-16 12:28:20 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
2018-04-16 12:28:20 0 [Note] /usr/sbin/mysqld (mysqld 5.6.39) starting as process 18545 ...
2018-04-16 12:28:20 18545 [Warning] Buffered warning: Changed limits: max_open_files: 1024 (requested 5000)
2018-04-16 12:28:20 18545 [Warning] Buffered warning: Changed limits: table_open_cache: 431 (requested 2000)
2018-04-16 12:28:20 18545 [Note] Plugin 'FEDERATED' is disabled.
/usr/sbin/mysqld: Unknown storage engine 'InnoDB'
2018-04-16 12:28:20 18545 [ERROR] Can't open the mysql.plugin table. Please run mysql_upgrade to create it. If I run "mysql_update", I get this: # mysql_upgrade Looking for 'mysql' as: mysql Looking for 'mysqlcheck' as: mysqlcheck Error: Failed while fetching Server version! Could be due to unauthorized access.
FATAL ERROR: Upgrade failed
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hive
04-16-2018
12:40 PM
Your initial indication helped: # wget http://repo.mysql.com/mysql-community-release-el7-5.noarch.rpm # sudo rpm -ivh mysql-community-release-el7-5.noarch.rpm But the below-given steps were leading to either a conflict or impossibility to execute "yum install mysql-community-release" from Ambari installation Wizard: # yum localinstall https://dev.mysql.com/get/mysql57-community-release-el7-8.noarch.rpm # yum install mysql-community-server # systemctl start mysqld.service
... View more
04-16-2018
10:28 AM
I followed your recommendations, but I go the following conflict error when installing MySQL server: resource_management.core.exceptions.ExecutionFailed: Execution of '/usr/bin/yum -d 0 -e 0 -y install mysql-community-release' returned 1. Error: mysql57-community-release conflicts with mysql-community-release-el7-7.noarch
You could try using --skip-broken to work around the problem
You could try running: rpm -Va --nofiles --nodigest
... View more
04-16-2018
10:13 AM
I solved this problem by opening the ports 2888 and 3888 that are used by ZooKeeper nodes for communicating between each other.
... View more
04-16-2018
09:54 AM
Sorry, yes, there is more content in LOG file. Please see the error messages below. There seem to be a problem with ZooKeeper, but I do not have any alerts for ZooKeeper. Zookeeper seems to run well: # jps -l | grep -i zookeeper 5043 org.apache.zookeeper.server.quorum.QuorumPeerMain # netstat -anp | grep 2181 tcp6 0 0 :::2181 :::* LISTEN 5043/java Errors log: 2018-04-16 09:16:23,821 ERROR resourcemanager.ResourceManager (LogAdapter.java:error(69)) - RECEIVED SIGNAL 15: SIGTERM 2018-04-16 09:16:25,315 INFO zookeeper.ClientCnxn (ClientCnxn.java:logStartConnect(1019)) - Opening socket connection to server eureambarislave1.local.eurecat.org/192.168.0.10:2181. Will not attempt to authenticate using SASL (unknown error) 2018-04-16 09:16:25,316 INFO zookeeper.ClientCnxn (ClientCnxn.java:primeConnection(864)) - Socket connection established, initiating session, client: /192.168.0.15:53808, server: eureambarislave1.local.eurecat.org/192.168.0.10:2181 2018-04-16 09:16:25,316 INFO zookeeper.ClientCnxn (ClientCnxn.java:run(1142)) - Unable to read additional data from server sessionid 0x0, likely server has closed socket, closing socket connection and attempting reconnect 2018-04-16 09:16:25,417 INFO recovery.ZKRMStateStore (ZKRMStateStore.java:runWithRetries(1227)) - Exception while executing a ZK operation.
org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /rmstore
at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
at org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore$1.run(ZKRMStateStore.java:326)
... View more
04-16-2018
09:41 AM
No, I do not have ResourceManager HA enabled. This is the content of "yarn-yarn-resourcemanager-eureambarislave2.local.eurecat.org.log": 2018-04-15 23:13:33,362 INFO resourcemanager.ResourceManager (LogAdapter.java:info(45)) - STARTUP_MSG:
/************************************************************ STARTUP_MSG: Starting ResourceManager STARTUP_MSG: user = yarn STARTUP_MSG: host = eureambarislave2.local.eurecat.org/192.168.0.15
STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.3.2.6.4.0-91 And this is the content of "yarn-yarn-resourcemanager-eureambarislave2.local.eurecat.org.out": SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
(-i) 64019 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 32768 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 8192 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
... View more
04-16-2018
09:27 AM
This is what I get: # jps -l | grep -i resourcemanager 4274 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager # netstat -anp | grep 8088
[root@eureambarislave2 sbin] # The netstat returns nothing.
... View more
04-16-2018
09:13 AM
By the way, I opened the port 8088 as follows: iptables -I INPUT 1 -p tcp --dport 8088 -j ACCEPT Is it correct?
... View more
04-16-2018
08:54 AM
I get the following output after running this command: root 811 31804 0 08:51 pts/0 00:00:00 grep --color=auto lsof -i :8088 | grep -i listen | awk '{print }' But the command "netstat -tunalp | grep LISTEN" does not show the port 8088.
... View more
04-15-2018
11:21 PM
I want to add Hive service to my cluster. In the Add Service Wizard I see the attached screen. It looks like MySQL is the default option. So, what is a recommended way to add Hive? I do not have any special requirement for selecting MySQL. My final goal is to install Spark2 which requires Hive.
... View more