Member since
02-29-2016
41
Posts
20
Kudos Received
0
Solutions
03-12-2018
06:02 AM
Is it possible to write into a location directly instead of hdfs path ?
... View more
07-25-2016
08:41 AM
Is there way to configure and send mail whenever hadoop environment is down, 1. Hive is down 2. One of data node is down or its space reached its threshold
... View more
Labels:
- Labels:
-
Apache Hadoop
04-25-2016
09:26 AM
@Benjamin Leonhardi All i need to do is on Hive external tables directly. 1.My above DDL statement was not working when i try to create parquet external table with snappy compression. 2. Is there a way to alter compression from snappy to ZIP in an existing hive external table.
... View more
04-22-2016
06:14 AM
1 Kudo
Throwing this error while creating Hive parquet table with snappy compression in hive beeline mode. Error: Error while compiling statement: FAILED: ParseException line 19:15 cannot recognize input near 'parquet' '.' 'compress' in table properties list (state=42000,code=40000) CREATE EXTERNAL TABLE testsnappy ( column bigint ) row format delimited fields terminated by ',' STORED as PARQUET LOCATION 'path' TBLPROPERTIES ("parquet.compress"="SNAPPY") " ; Also is there a way to set compression format for already created tables ?
... View more
Labels:
- Labels:
-
Apache Hive
03-18-2016
09:50 AM
Thanks much. Name node uses only RAM memory?..no space is required as such data node ?
... View more
03-17-2016
07:27 AM
2 Kudos
our HDFS cluster size is 16 TB. have 3 data nodes and 1 name node. 1.how to find out the storage size alloted for 3 data nodes ? 2. Does name node uses only RAM memory ? how to find its capacity as well
... View more
Labels:
- Labels:
-
Apache Hadoop
03-03-2016
11:57 PM
Yes Tim. That works. Here we need to join target table (table2). If CREATE SEQUENCE function is there in impala, it would be of more useful instead of doing these tricky things. Thanks for your reply.
... View more
03-02-2016
05:15 AM
Thanks Tim for answering it. I tried using UDFrowsequence java code and it always start from 1. http://svn.apache.org/repos/asf/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udf/UDFRowSequence.java Is there a way to start from max value by passing it ?
... View more
03-01-2016
01:01 AM
whenever a new record is inserted, sequence number has to be generated for key columns. first day : 100 record with sequence number from 1 to 100. second day : 10 new records. Sequence number should start from 101 to 110. We can achieve this in ETL tools. Also, In some databases, we have create sequence function and use the next value in select clause. Is it possible to achieve in impala ?
... View more
02-29-2016
11:44 PM
I want to generate sequence number for each new record in Impala. 1. First load - am loading 100 records with sequence number starting from 1 to 100. 2. Second load - i need to load new 10 records which are not present in first load . So my sequence number has to start from 101 to 110 for these new to records. I think there is no sequence function here and Next value cannot be used to hold the max value. Is this possible with any other function or UDFs has to be written ?
... View more
Labels:
- Labels:
-
Apache Impala