Member since
02-16-2016
176
Posts
197
Kudos Received
17
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3266 | 11-18-2016 08:48 PM | |
5822 | 08-23-2016 04:13 PM | |
1576 | 03-26-2016 12:01 PM | |
1480 | 03-15-2016 12:12 AM | |
15716 | 03-14-2016 10:54 PM |
03-14-2016
01:45 PM
4 Kudos
Does nifi HBase ClientService support connecting to Kerberized HBase ? Nifi - 0.4.0 HBase - 1.1.2
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache NiFi
03-14-2016
01:35 PM
Thank You @Artem Ervits for all the links. That was useful information. I am still struggling with getting sizing for my requirements. I will be moving less than 100Gb of data on a daily basis so volume is not a whole lot. But my data comes in spurts. 20Mbps rate at peak times will suffice. I am wondering if I can run my Nifi client from a VM with 8 Gb of RAM and 200Gb of disk space instead of investing in a server.
... View more
03-12-2016
11:26 PM
2 Kudos
Is there any guide available for NiFi capacity planning ? What load characteristic should be looked at ? - Number of concurrent processors running - Total throughput Any help is greatly appreciated.
... View more
Labels:
- Labels:
-
Apache NiFi
03-10-2016
05:47 PM
2 Kudos
@Gerd Koenig
In Ambari 2.0.1 port is hardcoded to 6080. Newer release of Ambari has a fix. https://github.com/apache/ambari/commit/f9e73665b48c44cb6e8118bb613d81584fddc497
... View more
03-10-2016
12:43 PM
4 Kudos
@Nelson KA Rajendran You need to create a dummy table with one row create table dummy(a string); insert into table dummy values ('a'); Then you can insert to your test_array table using INSERT INTO table test_array SELECT 1, array('a','b') from dummy; You can't insert a complex type directly to Hive table.
... View more
03-10-2016
10:51 AM
@Kuldeep Kulkarni Interesting question. You can probably enable Ranger Admin HA, then bring another node down. But this will still leave Usersync processes needing to be moved.
... View more
03-09-2016
06:22 PM
3 Kudos
Thank You @Artem Ervits @Sunile Manjee @Ancil McBarnett. I was able to get my solution using a combination of commands. hdfs dfs -cat $1 | head --bytes 10K > $SAMPLE_FILE
java -jar $AVRO_TOOLS_PATH/avro-tools-1.7.7.jar getschema $SAMPLE_FILE > $AVRO_SCHEMA_FILE
hdfs dfs -put $AVRO_SCHEMA_FILE $AVRO_SCHEMA_DIR head command needs to be used with --bytes option to get first 10K bytes. Then I used Avro tools to retrieve schema and copied schema back to HDFS.
... View more
03-09-2016
03:57 PM
Thanks @Ancil McBarnett @Sunile Manjee. I don't have .avsc file for schema. How can I extract Avro schema for this data ?
... View more
03-09-2016
03:42 PM
3 Kudos
I have a dataset that is almost 600GB in Avro format in HDFS. Whay is the most efficient way to create a Hive table directly on this dataset ? For smaller datasets, I can move my data to disk, use Avro tools to extract schema, upload schema to HDFS and create Hive table based on that schema. Is there a way to directly extract Avro schema from a dataset in HDFS without writing java code ?
... View more
Labels:
- Labels:
-
Apache Hive
03-07-2016
11:40 AM
1 Kudo
@Michael Dennis Uanang Can you please close this issue if it is resolved and open another one ? I am not clear on new issue. Are you able to write messages to topics but suddenly lose access because brokers are not registered with zookeepers ?
... View more