Member since
03-17-2016
132
Posts
106
Kudos Received
13
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2493 | 03-28-2019 11:16 AM | |
3103 | 03-28-2019 09:19 AM | |
2553 | 02-02-2017 07:52 AM | |
2682 | 10-03-2016 08:08 PM | |
1134 | 09-13-2016 08:00 PM |
10-06-2016
06:40 AM
Try to reset ambari admin password and have a try https://community.hortonworks.com/questions/449/how-to-reset-ambari-admin-password.html
... View more
10-03-2016
08:56 PM
@Logeswaran Radhakrishnan Please accept the answer if this worked for you
... View more
10-03-2016
08:08 PM
1 Kudo
You can can use simple hadoop -put source(local file system) destination (hdfs path). Make sure you have permission to access files on both local file system and hdfs.
... View more
10-03-2016
07:59 PM
2 Kudos
wrong mapreduce.jobhistory.address property's value. try to copy mapred-site.xml from the cluster
... View more
10-01-2016
12:23 PM
1 Kudo
Please accept the answer. So it will help to others.
... View more
09-13-2016
08:45 PM
Please run below command as hive user: /usr/bin/hive --service hiveserver2 --hiveconf hive.root.logger=DEBUG . yes kill hiveserver2 process on hadoop2 and run this on the same node.
... View more
09-13-2016
08:00 PM
7 Kudos
Configuring clients to use the High Availability feature The Atlas Web Service can be accessed in two ways: Using the Atlas Web UI: This is a browser based client that can be used to query the metadata stored in Atlas. Using the Atlas REST API: As Atlas exposes a RESTful API, one can use any standard REST client including libraries in other applications. In fact, Atlas ships with a client called AtlasClient that can be used as an example to build REST client access. In order to take advantage of the High Availability feature in the clients, there are two options possible.
HAProxy. Here is an example HAProxy configuration that can be used. Note this is provided for illustration only, and not as a recommended production configuration. For that, please refer to the HAProxy documentation for appropriate instructions. frontend atlas_fe
bind *:41000
default_backend atlas_be
backend atlas_be
mode http
option httpchk get /api/atlas/admin/status
http-check expect string ACTIVE
balance roundrobin
server host1_21000 host1:21000 check
server host2_21000 host2:21000 check backup
listen atlas
bind localhost:42000
The above configuration binds HAProxy to listen on port 41000 for incoming client connections. It then routes the connections to either of the hosts host1 or host2 depending on a HTTP status check. The status check is done using a HTTP GET on the REST URL /api/atlas/admin/status, and is deemed successful only if the HTTP response contains the string ACTIVE. Using automatic detection of active instance If one does not want to setup and manage a separate proxy, then the other option to use the High Availability feature is to build a client application that is capable of detecting status and retrying operations. In such a setting, the client application can be launched with the URLs of all Atlas Web Service instances that form the ensemble. The client should then call the REST URL /api/atlas/admin/status on each of these to determine which is the active instance. The response from the Active instance would be of the form {Status:ACTIVE}. Also, when the client faces any exceptions in the course of an operation, it should again determine which of the remaining URLs is active and retry the operation. The AtlasClient class that ships with Atlas can be used as an example client library that implements the logic for working with an ensemble and selecting the right Active server instance. Utilities in Atlas, like quick_start.py and import-hive.sh can be configured to run with multiple server URLs. When launched in this mode, the AtlasClient automatically selects and works with the current active instance. If a proxy is set up in between, then its address can be used when running quick_start.py or import-hive.sh.
... View more
09-13-2016
07:45 PM
Can you run this, So that we can exactly know whats gi\oing on /usr/bin/hive --service hiveserver2 --hiveconf hive.root.logger=DEBUG
... View more
09-13-2016
07:33 PM
4 Kudos
follow this http://blog.h2o.ai/2014/09/sparkling-water-tutorials/
... View more
09-13-2016
07:30 PM
5 Kudos
In order to provide HA for the index store, we recommend that Atlas be configured to use Solr as the backing index store for Titan. In order to configure Atlas to use Solr in HA mode, do the following:
Choose an existing SolrCloud cluster setup in HA mode to configure in Atlas (OR) Set up a new SolrCloud cluster.
Ensure Solr is brought up on at least 2 physical hosts for redundancy, and each host runs a Solr node. We recommend the number of replicas to be set to at least 2 for redundancy. Create the SolrCloud collections required by Atlas, as described in Installation Steps Refer to the Configuration page for the options to configure in atlas.properties to setup Atlas with Solr. Source - http://atlas.incubator.apache.org/HighAvailability.html
... View more