Member since
01-24-2017
8
Posts
4
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
908 | 02-14-2017 02:13 PM |
04-11-2017
07:56 AM
I don't get the philosophical part. Secondary index is orthogonal to what? The secondary index is an INTERSECTION of the primary key and what?
... View more
02-14-2017
02:13 PM
3 Kudos
3 or whatever number of slave cluster downtime is not really an exceptional case for replication. According to the documents linked below, HBase and Zookeeper will collect a backlog of edits and once the slave cluster is up again, replicate the older edits the same way as newer ones. So in this normal case, best way is to let it do its job. https://hbase.apache.org/0.94/replication.html#Normal_processing https://hbase.apache.org/0.94/replication.html#Non-responding_slave_clusters Abnormal cases can occur if table data gets corrupted and the replication breaks. Then you may have to copy. I don't know about the loads.
... View more
02-08-2017
03:36 PM
1 Kudo
As for >2 GB blobs, Hive STRING or even BINARY won't handle AFAIK. But that is just googled, Hive experts please add your thoughts. Please note that the "InvalidProtocolBufferException: Protocol message was too large. May be malicious. Use CodedInputStream.setSizeLimit() to increase the size limit." part in your stack trace tells you that you hit the limits of ProtocolBuffers, not Hive field type limitations. That could explain the 500 MB limit that you got in your investigations. In Hive code, orc input stream implementation I could see that there is 1 GB protobuf limit set but that is for the whole message and the blob is only a part of it.
... View more