Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1969 | 07-09-2019 12:53 AM | |
| 11880 | 06-23-2019 08:37 PM | |
| 9146 | 06-18-2019 11:28 PM | |
| 10133 | 05-23-2019 08:46 PM | |
| 4580 | 05-20-2019 01:14 AM |
09-09-2015
04:44 PM
The reader buffer size is indeed controlled by that property (io.file.buffer.size) but note that if you're doing short circuited reads then another property that also applies is (dfs.client.read.shortcircuit.buffer.size, 1 MB in bytes by default).
... View more
09-09-2015
04:16 PM
1 Kudo
You have installed the wrong Java JDK8 package. Please ensure to download the 64-bit JDK8 and remove your current 32-bit JDK8. A 64-bit JDK8 will print the below, for example, if you'd like to check and compare with your $JAVA_HOME/bin/java executable: ~ java -version java version "1.8.0_45" Java(TM) SE Runtime Environment (build 1.8.0_45-b14) Java HotSpot(TM) 64-Bit Server VM (build 25.45-b02, mixed mode)
... View more
09-08-2015
09:28 PM
Jobs typically read records - not entire blocks. Is your MR job doing anything different in this regard? Note that HDFS Readers do not read whole blocks of data at a time, and instead stream the data via a buffered read (64k-128k typically). That the block size is X MB does not translate into a memory requirement unless you are explicitly storing the entire block in memory when streaming the read.
... View more
09-04-2015
01:57 AM
Set the "DataNode Failed Volumes Tolerated" field in CM -> HDFS -> Configuration to the number of volumes each DN should tolerate the failure up to. XML property, if you do not use CM, is "dfs.datanode.failed.volumes.tolerated".
... View more
09-03-2015
05:51 PM
In spirit of https://xkcd.com/979/, feel free to mark the thread as resolved if it does help your cause, so others may find a solution quicker.
... View more
09-03-2015
05:49 PM
1 Kudo
Currently, the CM BDR feature does not carry any HBase replication abilities (we do support schedulable snapshot policies, but no replication/copies yet). You will need to utilise standard HBase techniques to copy over the data between your two clusters: http://blog.cloudera.com/blog/2013/11/approaches-to-backup-and-disaster-recovery-in-hbase/, and I'd recommend the ExportSnapshot method (if not live replication).
... View more
09-03-2015
05:42 PM
Indeed, as szehon mentions, the use of /root may be your problem, especially if you've invoked the Hive CLI via a sudo command. The /root is protected against access from anyone but the root user.
... View more
09-03-2015
05:16 PM
You will need the gateway copy, which exists under /etc/hive/conf/ on a Hive Gateway designated node (check Hive -> Instances in CM to find which hosts have a gateway role).
... View more
09-01-2015
10:18 PM
You can add a Java System Property setting that key, into the "Java Configuration Options for Zookeeper Server" field under ZooKeeper -> Configuration page. Add in the -D format, for ex. for 4 GiB, append: -Djute.maxbuffer=4294967296
... View more
08-25-2015
11:50 PM
You can use the FAST_DIFF encoding to perhaps, in one way, reduce the serialisation cost in HBase: http://archive.cloudera.com/cdh5/cdh/5/hbase/book.html#data.block.encoding.enable Also consider compressing your table - it will save a lot of space if you also make sure to use a proper HFile data block size (not the same as HDFS block size).
... View more