Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1968 | 07-09-2019 12:53 AM | |
| 11853 | 06-23-2019 08:37 PM | |
| 9135 | 06-18-2019 11:28 PM | |
| 10110 | 05-23-2019 08:46 PM | |
| 4569 | 05-20-2019 01:14 AM |
10-24-2016
03:18 AM
Harsh, Could you clarify answer for point no.1 1. Whether hbase snapshot make another copy of the table or not.
... View more
10-16-2016
10:13 AM
1 Kudo
Hi, I don't think that's possible given that most applications are based on HDFS semantics (strong consistency, POSIX compatible), and S3 simply isn't designed as a file system (eventual consistency, blob store). Plus, you lose data locality. As far as I know, most cloud use cases still use HDFS as temporary, intermediate storage, and use S3 as permanent, eventual storage. There've been several studies in using HDFS as meta store, and cloud as data store, but that's a huge work (see HDFS-9806) and probably in the Hadoop 4/CDH 7 timeframe. Hope this helps.
... View more
10-07-2016
01:44 AM
1 Kudo
Let's say your dataDir and old dataLogDir is /var/lib/zookeeper and now you're moving dataLogDir to /var/lib/zookeeper-log. First you change this in the service-wide configuration, which will make the stale configuration icon appear. Then you stop zk1, ssh into zk1 and run the following commands: $ mkdir -p /var/lib/zookeeper-log/version-2 $ cp /var/lib/zookeeper/version-2/log.* /var/lib/zookeeper-log/version-2/ $ chown -R zookeeper:zookeeper /var/lib/zookeeper-log Then you can start zk1 and wait until it's running and shows as either leader or follower in the Cloudera Manager service page too. After that's done, you can do the same with zk2 and finally with zk3 too. By this point the stale configuration alert should disappear and everything should be fine cluster-wide. As you said, the log.* files need to be copied only.
... View more
10-05-2016
12:57 AM
While it may appear possible to do this I'd strongly recommend against it because when you'd read back a written 150 MB MOB cell, it'd give you heap utilisation problems during the RPC encoding and transfer done by the RS. Its probably better to store the larger-than-10 MB files as HDFS files and store their paths in HBase.
... View more
10-05-2016
12:55 AM
You can find what operations are supported in the hdfs-fuse source: https://github.com/cloudera/hadoop-common/tree/cdh5.8.0-release/hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs (this is a directory-tree, look for the files with "impls" in their names for the syscalls supported). Git likely needs more advanced features from its used filesystem that HDFS currently does not offer. You can run git under strace to find out what that syscall is.
... View more
10-05-2016
12:49 AM
Cloudera offers Backup and Disaster Recovery (BDR) features as part of its enterprise offering that can do HDFS replication to other clusters, Hive metadata and data replication to other clusters, and also HBase snapshot backups to S3. This is documented in detail at https://www.cloudera.com/documentation/enterprise/latest/topics/cm_bdr_about.html Outside of this you can try to use DistCp for HDFS replication but for Hive replication you will need to manually propagate DDL-associated metadata.
... View more
10-05-2016
12:43 AM
1 Kudo
For (1), the answer right now is no. Once the dead node detection occurs, NameNode will swiftly act at re-replicating the identified lost replicas. There's something along the lines of what you need being worked upon upstream via https://issues.apache.org/jira/browse/HDFS-7877 but the work is still in progress and will only arrive in a future undetermined CDH release. For (2), you can hunt such files with replication factor of 1 and raise them to 2 and wait for under-replication count to reach 0 before you take the DN down. The change of replication factor is doable by the command 'hadoop fs -setrep'.
... View more
09-20-2016
06:55 AM
1 Kudo
Yes, you need to switch Oozie to submit over YARN and not MRv1. The switching guide covers this aspect.
... View more
09-20-2016
06:00 AM
can you tell me how to connect through java api to hbase cluster ( 4 node cluster running in VM's). from windows to Distribution running in server vm/s
... View more
09-12-2016
01:04 PM
1 Kudo
Hi harsh, The issue was with JCE files, i updated them in the wrong location instead of /usr/java/jdk1.7.0_25/jre/lib/security/. Once i have updated the JCE files in the above location, now i'm able to access HDFS. Thanks for the Help! Regards, Cibi
... View more