Member since
05-28-2015
47
Posts
28
Kudos Received
7
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5901 | 06-20-2016 04:00 PM | |
10604 | 01-16-2016 03:15 PM | |
11181 | 01-16-2016 05:06 AM | |
5173 | 01-14-2016 06:45 PM | |
2915 | 01-14-2016 01:56 AM |
11-06-2024
12:29 AM
1 Kudo
Hi, I am trying to connect hive database from nodejs using hive-driver npm. In that code,the session cannot be established to access the hive database. I am putting the console for each and every variables,but the log was printed,before the session establishing code only.please anyone help me on this.please refer a snap shot below.
... View more
07-29-2020
01:15 PM
Apache Hive Strengths: The Apache Hive encourages questioning and overseeing huge datasets living in circulated capacity. Based on head of Apache Hadoop, it gives: Tools to empower simple data separate/change/load (ETL) A system to force structure on an assortment of data positions Access to documents put away either legitimately in Apache HDFS or in other data stockpiling frameworks, for example, Apache HBase Query execution by means of MapReduce Hive characterizes a straightforward SQL-like inquiry language, called QL, that empowers clients acquainted with SQL to question the data. Simultaneously, this language additionally permits developers who know about the MapReduce system to have the option to connect their custom mappers and reducers to perform increasingly modern investigation that may not be bolstered by the inherent capacities of the language. QL can likewise be stretched out with custom scalar capacities (UDF's), accumulations (UDAF's), and table capacities (UDTF's). Ordering to give quickening, list type including compaction and Bitmap file as of 0.10. Diverse capacity types, for example, plain content, RCFile, HBase, ORC, and others. Metadata stockpiling in a RDBMS, essentially decreasing an opportunity to perform semantic checks during inquiry execution. Working on compacted data put away into the Hadoop biological system utilizing calculations including DEFLATE, BWT, smart, and so on. Worked in client characterized capacities (UDFs) to control dates, strings, and other data-mining tools. Hive underpins stretching out the UDF set to deal with use-cases not bolstered by worked in capacities. SQL-like questions (HiveQL), which are verifiably changed over into MapReduce, or Spark employments. Apache Spark Strengths: Flash SQL has various intriguing highlights: it underpins various document arrangements, for example, Parquet, Avro, Text, JSON, ORC it bolsters data put away in HDFS, Apache HBase, Cassandra and Amazon S3 it underpins traditional Hadoop codecs, for example, smart, lzo, gzip it gives security through authentification by means of the utilization of a "common mystery" (spark.authenticate=true on YARN, or spark.authenticate.secret on all hubs if not YARN) encryption, Spark underpins SSL for Akka and HTTP conventions it bolsters UDFs it bolsters simultaneous questions and deals with the distribution of memory to the employments (it is conceivable to indicate the capacity of RDD like in-memory just, circle just or memory and plate it underpins reserving data in memory utilizing a SchemaRDD columnar arrangement (cacheTable(""))exposing ByteBuffer, it can likewise utilize memory-just storing uncovering User object it underpins settled structures When to utilize Spark or Hive- Hive is as yet an extraordinary decision when low inactivity/multiuser support isn't a prerequisite, for example, for clump preparing/ETL. Hive-on-Spark will limit the time windows required for such handling, yet not to a degree that makes Hive appropriate for BI Flash SQL, lets Spark clients specifically use SQL builds when composing Spark pipelines. It isn't proposed to be a universally useful SQL layer for intelligent/exploratory investigation. In any case, Spark SQL reuses the Hive frontend and metastore, giving you full similarity with existing Hive data, questions, and UDFs. Flash SQL incorporates a cost-based streamlining agent, columnar capacity and code age to make inquiries quick. Simultaneously, it scales to a great many hubs and multi hour inquiries utilizing the Spark motor, which gives full mid-question adaptation to internal failure. The exhibition is greatest bit of leeway of Spark SQL.
... View more
05-06-2020
03:24 AM
This could be permission issue. you can see the hive server2 log for the error. Log will be in /var/log/hive on the node to which you connect the hive
... View more
12-16-2019
04:59 AM
@rbanerjee, to answer your question, yes, temporary tables are independent and isolated between sessions for the same user. Here, https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-TemporaryTables, it says "A table that has been created as a temporary table will only be visible to the current session" I have verified that. As the same user I created a temporary table of the same name in two concurrent Beeline sessions. They are independent, as confirmed by INSERT/SELECT statements and conclusively by their locations in HDFS: Session 1: SHOW CREATE TABLE... LOCATION ... /tmp/hive/hive/4d4e879d-5806-425e-9051-d764b117c71d/_tmp_space.db/d3b5f635-1251-4cda-be59-79555c8643fd Session 2: SHOW CREATE TABLE... LOCATION .. /tmp/hive/hive/15049889-fc99-4890-b3eb-0a5d22c73083/_tmp_space.db/ac5a2d3c-a61f-4797-afa4-c88940a10147
... View more
11-08-2018
08:39 AM
@Muhammad Taimoor Yousaf You'd need to have a valid ticket in in the machine you are launching pyspark shell (which is test-m1). Looking at the exception it is clearly saying 'No valid credentials provided'. Hope this helps.
... View more
06-27-2016
03:43 PM
@mqureshi Thanks for your response. yes it is quite a custom requirement. I thought its better to check with the community if anyone has implemented this kinda stuff. I am trying to use either hadoop custom input format or python UDF's to get this done. There seems to be no straightforward way of doing this in spark. I can not use spark pivot also as it supports only column as of now right?.
... View more
01-18-2016
05:52 AM
Thank you Neeraj Sabharwal, yes, this is my fault not to go through the documentation first. i do first check documentations before touching any technology and i got this habbit from oracle. i dont know why in Hortonworks i skip documentations, may be it is new for me and a bit difficult to find the proper documentation... i dont know. but i promise, i will check them from now on. i really appriciate your kind replies and support. thank you so much.
... View more
12-29-2015
01:28 AM
Thank Chris
... View more
12-29-2015
01:25 AM
Thanks Chris!!
... View more
12-30-2015
06:27 PM
@Gangadhar Kadam As a best practice, please accept the answer if you are satisfied with answer. Then, we can close this question.
... View more