Member since
05-18-2016
71
Posts
39
Kudos Received
6
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1747 | 12-16-2016 06:12 PM | |
645 | 11-02-2016 05:35 PM | |
3447 | 10-06-2016 04:32 PM | |
1022 | 10-06-2016 04:21 PM | |
1083 | 09-12-2016 05:16 PM |
03-07-2018
10:51 PM
try adding this parameter to your sqoop job -D sqoop.export.records.per.statement=100 you can micro batch your transactions.
... View more
12-22-2017
06:27 PM
if you are using 2.6 version or later, you can turn on ACID and execute delete commands. you can audit your delete via your app if it is needed. Otherwise, delete, merge and update can run on Hive directly with ACID.
... View more
05-16-2017
02:36 PM
by Inserting into table using the construct INSERT INTO SELECT FROM... you are essentially bacthing/microbatching the data. You would have to create scripts/code to handle exceptions as you are loading. There are multiple ways of doing this. But there is no automated way unless you use ETL/ELT tools.
... View more
05-15-2017
06:35 PM
How are you loading the table? Are you loading them from CSV files, or using sqoop. Each of those ways you could enable logging at job level to track changes and restart the batch. If you have written your own scripting, on error you should try and write those files to error directories so you could go back and look for errors, fix them and reload them.
... View more
04-17-2017
01:24 AM
This works perfectly with Field Cloud. If you want to run some queries on phoenix by following this and Phoenix and Hbase tutorials this is an awesome demoable material
... View more
04-03-2017
06:27 PM
Also, how big is data in your table.. are you doing some sort of Limit or where clause when you run the query?
... View more
04-03-2017
06:18 PM
Are these tables External Tables? In the case of external tables you would have manually clean the folders by removing the files and folders that are referenced by the table ( using hadoop fs -rm command)
... View more
02-03-2017
07:17 PM
1 Kudo
Is this an external table? if it is an external table then write a perl/shell script that runs daily based on a pattern it will remove files older than a certain date. If it is not, and if it is a managed table with ORC/Parquet format, then it depends on how you load the table. While loading the table you could setup logic to populate the date in the table partition by that date and as @Avijeet Dash mentioned drop the partition that is 7 days old.
... View more
12-16-2016
06:12 PM
1 Kudo
Ambari should always run on 8080 port, so connecting to http://localhost:8080 should take you directly to the ambari login. Zeppelin runs on port 9995.
... View more
12-15-2016
09:07 PM
sudo su - hdfs, then execute your commands. the xdl3 user does not have write access to /xdl/tmp directory. Also i hope you dont have any acls setup.
... View more