Member since
02-10-2016
36
Posts
14
Kudos Received
0
Solutions
02-17-2016
04:34 PM
1 Kudo
I'm using Hadoop 2.6.
... View more
02-17-2016
01:08 PM
1 Kudo
Thanks for the response. Yes, I'm able to access S3 through simple Hive queries.From the logs, I could see that the map-reduce job is trying to connect to "hive-bucket.s3.amazonaws.com:443", which doesn't exist. I need to connect to a custom S3 endpoint, which is "s3-customlocation.net". I have gone through the hdfs-site configuration,but I couldnt find any parameter to set custom endpoint.
... View more
02-17-2016
12:41 PM
1 Kudo
I'm using a custom S3 for Eucalyptus, not the AWS one. I have been trying to resolve this since past few weeks.
... View more
02-17-2016
12:35 PM
1 Kudo
I have a Hadoop cluster(HDP 2.2) set-up in Eucalyptus environment. I have created an external table in Hive(0.14), using the below query: CREATE EXTERNAL TABLE tempbatting (col_value STRING) LOCATION 's3n://hive-bucket/';
I'm using a custom S3 location, so I have set jets3t property in Hive configuration directory as below: set s3service.https-only = true;
set s3service.s3-endpoint = s3-customlocation.net;
set s3service.s3-endpoint-http-port = 80;
set s3service.s3-endpoint-https-port = 443;
set s3service.disable-dns-buckets = true;
set s3service.enable-storage-classes = false;
Though I'm able to execute simple select queries on the table successfully, the aggregate queries are failing. Below are the logs: Error: java.io.IOException: java.lang.reflect.InvocationTargetException
at org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97)
at org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57)
Caused by: org.apache.http.conn.ConnectTimeoutException: Connect to hive-bucket.s3.amazonaws.com:443 timed out
at org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:416)
From the logs, the map-reduce job seems to access Amazon S3. I have tried using the set command for Hive(set fs.s3n.endpoint=s3-customlocation.net), but it didn't seem to work. Is there a way to specify custom end-point?
... View more
Labels:
- Labels:
-
Apache Hive
02-11-2016
12:26 PM
1 Kudo
Noted. Thank you.
... View more
02-11-2016
11:53 AM
1 Kudo
Setting no proxy for the fqdn in /etc/profile solved the issue. Thanks.
... View more
02-11-2016
11:24 AM
1 Kudo
The JDK version is 1.6.0
... View more
02-11-2016
11:11 AM
1 Kudo
Yes, it says connected.
... View more
02-11-2016
11:03 AM
1 Kudo
Thanks for the response. I have set no proxy for local addresses in the host machine using export no_proxy from the command line. Below is the command that fails while starting the service: resource_management.core.exceptions.Fail: Execution of 'curl -sS -L -w '%{http_code}' -X GET 'http://fqdn:50070/webhdfs/v1/app-logs?op=GETFILESTATUS&user.name=hdfs'' returned status_code=502.
... View more
02-11-2016
06:20 AM
2 Kudos
I'm trying to set up a single-node Hadoop cluster (HDP2.3 version) using
Ambari 2.2 in Eucalyptus environment. I have used Private IP address as recommended while
registering the VM's (RHEL 6.5) with Ambari. When I tried to bring up Hadoop services (MR2, Hive,
Yarn) , I'm getting the below error: <LI id=L_11001_11>Error Code 11001: Host not found
<LI id=L_11001_12>Background: This error indicates that
the gateway could not find the IP address of the website you are trying to
access. This is usually due to a DNS-related error.
<LI id=L_11001_13>Date: 2/9/2016 5:57:11 AM [GMT]
<LI id=L_11001_14>Server: FIESPRX004.xxx.net
<LI id=L_11001_15>Source: DNS error
</UL>
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
- « Previous
-
- 1
- 2
- Next »