Member since
08-05-2016
76
Posts
10
Kudos Received
13
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2664 | 11-06-2018 08:19 PM | |
1886 | 08-31-2018 05:34 PM | |
1425 | 05-02-2018 02:21 PM | |
2075 | 04-28-2018 09:32 PM | |
2282 | 11-09-2017 06:02 PM |
12-05-2017
10:50 PM
can you create a new issue and attach the logs. it is hard to see what is going on without logs. One way to check this is to first issue an explain query that will show you the druid query then you can copy that query and try it your self via curl command.
... View more
11-30-2017
06:00 PM
1 Kudo
FYI Derby is a local instance DB used only for testing. For production please use Mysql or Postgres.
... View more
11-09-2017
06:08 PM
Concerning the DIY I cannot provide an exact answer but it is more a case by case I guess. From Druid perspective generally speaking Druid only uses Ambari to manage configs, therefore, it is doable to drop in some jars or update the HDP package only but this needs some level of testing to make sure nothing is breaking between versions.
... View more
11-09-2017
06:02 PM
current distribution HDP 2.6.3 and Ambari 2.6 has druid 0.10.1 plus tons of patches.
... View more
10-23-2017
01:23 PM
@Roshan Dissanayake 1- keep in mind that the indexes have some extra overhead thus technically speaking some part of the data will be replicated but in a form of index thus compressed and more concise. 2- Hive will not manage the lifecycle of druid indexes you need to setup some Ozie (or any another workflow manager) to do the create table / insert into statements or drop table to keep the indexes up to date. 3- on a side not sure how the updates lands in your hive system, but if your pattern is mostly append/insert over a period of time then druid is designed for that usecase since data will be partition using time column.
... View more
10-18-2017
02:11 PM
Thanks yes we will update the Wiki. Am not 100%, but I hope this will make it to 2.6.3, we are pushing it to the finish line. Thanks.
... View more
10-18-2017
01:53 PM
1 Kudo
Sorry currently we only issue Time-series and Group-by queries and it is by design. The reason we backed of TopN is that Druid's TopN algorithm is an approximate thus is not the exact result, that is why we use Group-by all the time. Since we are a SQL shop, we need to make sure that results are correct rather than running fast. Although we are adding a new feature to allow approximate results thus the CBO will use TopN when possible if the approximate flag is turned On. Thanks!
... View more
10-12-2017
05:47 PM
1 Kudo
Hi @Roshan Dissanayake The integration is production ready, we are planning on GA version HDP 2.6.3 which going to be released soon. To answer your question about performance, i don't think the data size is an issue since Druid/LLAP can scale horizontally. The real question is how much of your query can be pushed to the druid cluster. This might require rethinking the schema of the OLAP Cubes and maybe rewrite some of the queries. I will be happy to help you with that if you can share the queries and schema.
... View more
10-11-2017
02:17 PM
Not sure what is the final use case, but one way to do this is: hive> CREATE TABLE foo (bar CHAR(8));
hive> insert into foo values ("00008DAC");
hive> select * from foo;
OK
00008DAC
... View more
10-04-2017
11:00 PM
can you start a new thread and add more information about the install process and the stack trace ?
... View more