Member since
09-24-2015
527
Posts
136
Kudos Received
19
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2890 | 06-30-2017 03:15 PM | |
| 4375 | 10-14-2016 10:08 AM | |
| 9598 | 09-07-2016 06:04 AM | |
| 11634 | 08-26-2016 11:27 AM | |
| 1912 | 08-23-2016 02:09 PM |
12-28-2016
09:00 PM
1 Kudo
@Roberto Sancho Are you running HDP or CDH? What version? Which tutorials? Please provide links to tutorial. Check your libraries versions and alignment. Also, your script seems to pass a null value where a value is expected. Validate input.
... View more
07-01-2016
03:47 PM
60 values of CODNRBEENF per day or in total? If you have 60 unique CODNRBEENF per day, leading with that column would be better. Otherwise, the date is probably better over time. If you are also querying on CODINTERNO and CODTXF (with FECHAOPRCNF and CODNRBEENF), then it makes sense to include them. It is not a problem to have four columns in the primary key constraint.
... View more
06-22-2016
11:08 PM
Can you try put the below first. load 'hbase://table/JOURNEY_OFICINA_HBASE' using org.apache.phoenix.pig.PhoenixHBaseLoader(zkQuorum);
... View more
06-22-2016
07:17 PM
Could you please explain the use case you have? Do you plan to query this data using HBase or Phoenix? In Phoenix case you just use the regular sql statements via jdbc driver. For HBase you need to handle everything yourself. So, looking for a specific records you need to run something like get 'JOURNEY_OFICINA_HBASE', '01982016-06-01 00:00:00 ' You need those trailing whitespaces since you are using the fixed size types. So the whole length of the rowkey should be exactly 25 symbols.
... View more
06-21-2016
06:40 PM
Ranger and Knox are complimentary. Knox is for perimeter security. It allows you to control the entry point for users to your cluster. You can put Knox behind a load balancer and shield the users from access to specific servers in the cluster. Ranger and Knox integrate well together, so you can use Ranger to grant permissions to users for Knox. This tutorial can walk you through setting up Knox: http://hortonworks.com/hadoop-tutorial/securing-hadoop-infrastructure-apache-knox/ And here is some good info on Knox: http://hortonworks.com/apache/knox-gateway/
... View more
06-17-2016
06:56 PM
is drill client configured for Acid? If you see the change in Hive, you must have set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager and hive.support.concurrency=true. Are the same properties set for the session through which Drill connects to Hive?
... View more
06-17-2016
01:07 PM
2 Kudos
@Roberto Sancho Hope below post help you to understand the differences. http://stackoverflow.com/questions/13715044/hive-cluster-by-vs-order-by-vs-sort-by
... View more
06-16-2016
08:54 PM
Agree with Sindhus Comments. The link provides some basic setup to optimize the queries on Hive. @Roberto Sancho Can you please let us know how is the table partitioned, bucketed. If it is partitioned, can you please give us if the where clause makes use of the partition columns.
... View more
06-14-2016
06:08 PM
it work now like this: first create the table from command line: /usr/hdp/current/phoenix-client/bin
./sqlline.py zookeeperHostname
create table test (mykey integer not null primary key, mycolumn varchar);
upsert into test values (1,'Hello');
upsert into test values (2,'World!');
select * from test;
and from RStudio like this:
conn <- dbConnect(drv,"jdbc:phoenix:lnxbig04.cajarural.gcr,lnxbig05.cajarural.gcr,lnxbig06.cajarural.gcr:2181:/hbase-unsecure")
query <- paste("select * from test")
B <- dbGetQuery(conn, query)
dbListTables(conn)
... View more