Member since
05-16-2016
785
Posts
114
Kudos Received
39
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2328 | 06-12-2019 09:27 AM | |
| 3579 | 05-27-2019 08:29 AM | |
| 5728 | 05-27-2018 08:49 AM | |
| 5244 | 05-05-2018 10:47 PM | |
| 3113 | 05-05-2018 07:32 AM |
09-11-2017
04:49 AM
1 Kudo
@syamsri Ok Could you please let ,me know the file format that you are using for Hive ( testTableNew ) , Hive supports Delete Update only on ORC format starting from 0.14 . Try creating a table with ORC format , if you want more flexibility then try Apache KUDU but it has it owns merits and demerits . Hope this helps . CREATE TABLE Sample (
id int,
name string
)
CLUSTERED BY (id) INTO 2 BUCKETS STORED AS ORC
TBLPROPERTIES ("transactional"="true",
"compactor.mapreduce.map.memory.mb"="2048",
"compactorthreshold.hive.compactor.delta.num.threshold"="4",
"compactorthreshold.hive.compactor.delta.pct.threshold"="0.5"
);
... View more
09-09-2017
09:06 PM
1 Kudo
@manuspark3 @SupriyaPS If scale is not specified, it defaults to 0 . If no precision is supplied , it defaults to 10. Yes decimal works without precison.
... View more
09-08-2017
09:24 AM
When ever we use ORC we go with String since it has vectorization which gives a , extra kick try the below query sometimes copy / paste will cause issue because of the extra space that comes along with sometimes . so type if you could Method 1 set hive.enforce.bucketing = true; CREATE TABLE TEST (id DECIMAL,name STRING) CLUSTERED BY (id) INTO 2 BUCKETS STORED AS ORC TBLPROPERTIES("transactional"="true"); Note - I did not test query in my local vm please let me know if this helps .
... View more
09-06-2017
01:08 AM
1 Kudo
if you have hive shell or beeline - you can excute the same code nothing different or you can try hue web ui to export the hive results as .csv althought hue is not that good in downloading big tables .
... View more
09-04-2017
05:39 PM
I am glad , did you fix your ip /hostname in /etc/hosts file ?
... View more
09-03-2017
02:36 AM
1 Kudo
Apache Nifi is more of data routing and transformation tool outside the Hadoop cluster ,it has its own JVM but needs hadoop configuration if you are trying to push data to hdfs. It has many built processor that comes handy for pushing data quickly , nifi community is little big Streamsets is similar , you can deploy inside hadoop cluster and manage using Cloduera manager Apache kafak is again a streaming platform more of real time data just NIFI but if you want to do any aggregation on the fly Apache flume - Flume is also disturbing platform has concept of source - channel - sink . we mostly use to push logs to hdfs ,hbase has many built-in source ,sink and you can create custom one just like nifi processor if you looking for alternative for Apache Nifi you can use Apache flume, Apache kafaka , streamsets. Each one has its own use case though. Cloudera Navigator - Is used for Data management and security under one hood for hadoop platform
... View more
09-02-2017
06:47 PM
Change your host file / hostname 192.168.200.21 server1
# The following lines are desirable for IPv6 capable hosts
::1 ip6-localhost ip6-loopback
fe00::0 ip6-localnet
ff00::0 ip6-mcastprefix
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters also change the hostname i am not sure the path in Ubuntu it should be under /etc/hostname -> server1 restart the network Also do an echo $hostname or $HOSTNAME to see if it is reflecting Finally restart all the dameons . let me know if that helps
... View more
09-01-2017
03:45 AM
whats your CDH / Hive version ? could you let me know whether you had set the below properties if not set and could you let me know if that helps . set hive.exec.reducers.bytes.per.reducer=<number>
set hive.exec.reducers.max=<number>
set mapreduce.job.reduces=<number>
... View more
08-31-2017
05:54 AM
Could you please share the logs using the "Insert code " tool in the toolbar . beside the bold, italic, underline spoiler tag in the reply text area.
... View more
08-30-2017
01:54 AM
@josholsan Sure thing 🙂
... View more