Member since
08-14-2017
8
Posts
0
Kudos Received
0
Solutions
10-23-2017
01:23 PM
@Roshan Dissanayake 1- keep in mind that the indexes have some extra overhead thus technically speaking some part of the data will be replicated but in a form of index thus compressed and more concise. 2- Hive will not manage the lifecycle of druid indexes you need to setup some Ozie (or any another workflow manager) to do the create table / insert into statements or drop table to keep the indexes up to date. 3- on a side not sure how the updates lands in your hive system, but if your pattern is mostly append/insert over a period of time then druid is designed for that usecase since data will be partition using time column.
... View more
09-20-2017
01:33 PM
If you want to have a default value of "null" then the type of your field needs to be a union of null and the real type. For example, for timestamp you would need: "type": ["long", "null"]
... View more
01-28-2019
11:09 AM
@Greg Keys I am having the same issue with hdf 3.3.1. I have checked the schema file and the input file as well. I have done what was mentioned by @Sriharsha Chintalapani Schema {
"namespace": "hdf.heaptrace.com",
"type": "record",
"name": "PatientField",
"fields": [
{"name": "Patient_name","type": "string"}
]
} JSON data {"Patient_name":"john"}
Please help !!! I have converted data from json to avro and then back again as well using avro tools.
... View more