Member since
05-30-2018
1322
Posts
715
Kudos Received
148
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 4038 | 08-20-2018 08:26 PM | |
| 1936 | 08-15-2018 01:59 PM | |
| 2369 | 08-13-2018 02:20 PM | |
| 4100 | 07-23-2018 04:37 PM | |
| 5005 | 07-19-2018 12:52 PM |
09-29-2016
04:35 AM
I found the issue. This is a bug. Engineering working on issue.
... View more
09-28-2016
03:55 AM
@Sunile Manjee thanks!
... View more
09-27-2016
04:59 AM
@nmaillard & @Randy Gelhausen great stuff. thank you
... View more
10-31-2016
04:09 PM
4 Kudos
@Sunile Manjee As I mentioned in my original response: Not without Livy. Yes with Livy (@vshukla). However, it is exposed only to Zeppelin, for now.
... View more
09-20-2016
07:15 PM
3 Kudos
In the HDP 2.5 release notes it says that Hive 2.1 is TP http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.0/bk_release-notes/content/tech_previews.html Hive New in this release: Stored Procedures Client-Side LLAP Daemons (HIVE-7193) LLAP / Spark security interface Hive 2.1 In the HDP 2.5 Release email to customers it was stated that Hive 2.1 is TP but that Hive ACID is certified for production with Hive 1.2.1
Apache Hive
Includes Apache Hive 1.2.1 for production and Hive 2.1 (Technical Preview) for cutting-edge performance Hive LLAP (Technical Preview): Persistent query servers and optimized in-memory caching for blazing fast SQL. Up to 25x faster for BI workloads. 100% compatible with existing Hive workloads Hive ACID and Streaming Ingest certified for production use with Hive 1.2.1 Dynamic user-based security policies for data masking and filtering HPL/SQL: Procedural programming within Hive Hive View v1.5.0, improved robustness and security Parquet format fully certified with Hive 1.2.1 / 2.1 In the Hortonworks.com Hive Overview section it states (confusingly) that ACID is GA in Hive 2.1 (though originated in 0.14) http://hortonworks.com/apache/hive/#section_3
... View more
09-26-2016
06:01 PM
@Rajkumar Singh thanks. good stuff.
... View more
09-19-2016
02:18 AM
Assuming that I installed four instances of PostgreSQL 9.3, and what if, for example, the Ranger database fails, it means failure also for HDFS and Hive security (among others). So these components (Ambari, Hive, Oozie, Ranger) are not entirely independent to warrant that a failure in their respective databases means the other will be operating smoothly. Someone suggested to me to do a single database instance for all four of the services in High Availability mode (master-slave, with warm-standby) or in multi-node cluster, four separate database instances (same distro and version presumably) in High Availability. Although for inexperienced DB admin (like me), this is quite a chore. As I have read from PostgreSQL documentation, there are a number of solutions for High Availability mode, like Shared Disk Failover, Transaction Log Shipping, etc. What solution did you employ for PostgreSQL HA. Can those who have done this in production cluster share how you did this? @Sunile Manjee
... View more
09-18-2016
08:18 PM
1 Kudo
https://developer.ibm.com/hadoop/2016/04/26/submit-spark-jobs-via-rest-iop-4-1-using-livy/ http://livy.io/overview.html http://www.slideshare.net/JenAman/livy-a-rest-web-service-for-apache-spark https://spark-summit.org/eu-2015/events/building-a-rest-job-server-for-interactive-spark-as-a-service/ Livy with Zeppelin
http://www.slideshare.net/HadoopSummit/apache-zeppelin-livy-bringing-multi-tenancy-to-interactive-data-analysis
https://github.com/apache/zeppelin/blob/master/docs/interpreter/livy.md
Using Livy with Zeppelin
https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.0/bk_zeppelin-component-guide/content/ch_stop-services.html
https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.0/bk_zeppelin-component-guide/content/install-livy.html
... View more
12-09-2016
08:38 PM
1 Kudo
Atlas HBase is made HA by configuring as distributed HBase (vs default standalone) http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.0/bk_data-governance/content/ch_hdp_data_governance_install_atlas_ambari.html
... View more
09-13-2016
08:00 PM
7 Kudos
Configuring clients to use the High Availability feature The Atlas Web Service can be accessed in two ways: Using the Atlas Web UI: This is a browser based client that can be used to query the metadata stored in Atlas. Using the Atlas REST API: As Atlas exposes a RESTful API, one can use any standard REST client including libraries in other applications. In fact, Atlas ships with a client called AtlasClient that can be used as an example to build REST client access. In order to take advantage of the High Availability feature in the clients, there are two options possible.
HAProxy. Here is an example HAProxy configuration that can be used. Note this is provided for illustration only, and not as a recommended production configuration. For that, please refer to the HAProxy documentation for appropriate instructions. frontend atlas_fe
bind *:41000
default_backend atlas_be
backend atlas_be
mode http
option httpchk get /api/atlas/admin/status
http-check expect string ACTIVE
balance roundrobin
server host1_21000 host1:21000 check
server host2_21000 host2:21000 check backup
listen atlas
bind localhost:42000
The above configuration binds HAProxy to listen on port 41000 for incoming client connections. It then routes the connections to either of the hosts host1 or host2 depending on a HTTP status check. The status check is done using a HTTP GET on the REST URL /api/atlas/admin/status, and is deemed successful only if the HTTP response contains the string ACTIVE. Using automatic detection of active instance If one does not want to setup and manage a separate proxy, then the other option to use the High Availability feature is to build a client application that is capable of detecting status and retrying operations. In such a setting, the client application can be launched with the URLs of all Atlas Web Service instances that form the ensemble. The client should then call the REST URL /api/atlas/admin/status on each of these to determine which is the active instance. The response from the Active instance would be of the form {Status:ACTIVE}. Also, when the client faces any exceptions in the course of an operation, it should again determine which of the remaining URLs is active and retry the operation. The AtlasClient class that ships with Atlas can be used as an example client library that implements the logic for working with an ensemble and selecting the right Active server instance. Utilities in Atlas, like quick_start.py and import-hive.sh can be configured to run with multiple server URLs. When launched in this mode, the AtlasClient automatically selects and works with the current active instance. If a proxy is set up in between, then its address can be used when running quick_start.py or import-hive.sh.
... View more