Member since
08-08-2013
339
Posts
132
Kudos Received
27
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
14591 | 01-18-2018 08:38 AM | |
1497 | 05-11-2017 06:50 PM | |
8963 | 04-28-2017 11:00 AM | |
3357 | 04-12-2017 01:36 AM | |
2769 | 02-14-2017 05:11 AM |
01-06-2014
01:34 AM
I'd also highly recommend the usage of the Quorum Journal. In addition to the requirement of having additional NFS infrastructure, NFS itself isn't that reliable and we had several issues with its availability in the past.
... View more
01-06-2014
01:30 AM
Hi, many thanks for your explanation. I'll check out the custom sink wrapper stuff...
... View more
12-15-2013
11:42 PM
Hi Eric, many thanks for answering that quickly. Maybe I have to check IntelliJ as alternative to the heavyweight-Eclipse 😉 br...Gerd...
... View more
12-14-2013
12:39 PM
Hi, I recently tried to create an Eclipse project of the kite-sdk, but the step to create the "Kite Data HBase module" failed with this error: " [ERROR] Failed to execute goal on project kite-data-hbase: Could not resolve dependencies for project org.kitesdk:kite-data-hbase:jar:0.10.1-SNAPSHOT: Could not find artifact org.kitesdk:kite-data-core:jar:0.10.1-SNAPSHOT in com.cloudera.releases (https://repository.cloudera.com/artifactory/cloudera-repos/) -> [Help 1] " Yes, the given URL doesn't contain com.cloudera.releases... The steps I executed are: git clone https://github.com/kite-sdk/kite.git cd git export JAVA_HOME=/usr/lib/jvm/java-6-openjdk-amd64/jre && mvn eclipse:eclipse The output looks like: ... [INFO] ------------------------------------------------------------------------ [INFO] Building Kite Data HBase Module 0.10.1-SNAPSHOT [INFO] ------------------------------------------------------------------------ [INFO] [INFO] >>> maven-eclipse-plugin:2.9:eclipse (default-cli) @ kite-data-hbase >>> [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Kite Development Kit .............................. SUCCESS [0.838s] [INFO] Kite Data Module .................................. SUCCESS [0.013s] [INFO] Kite Data Core Module ............................. SUCCESS [2.090s] [INFO] Kite Data Crunch Module ........................... SUCCESS [0.621s] [INFO] Kite Data Flume Module ............................ SUCCESS [0.280s] [INFO] Kite Data Hive Module ............................. SUCCESS [1.087s] [INFO] Kite Data HBase Module ............................ FAILURE [0.339s] [INFO] Kite Flume Avro Event Serializer .................. SKIPPED [INFO] Kite Maven Plugin ................................. SKIPPED [INFO] Kite Tools Module ................................. SKIPPED [INFO] Kite Morphlines Parent ............................ SKIPPED [INFO] Kite Morphlines Core .............................. SKIPPED [INFO] Kite Morphlines Avro .............................. SKIPPED [INFO] Kite Morphlines JSON .............................. SKIPPED [INFO] Kite Morphlines Maxmind ........................... SKIPPED [INFO] Kite Morphlines Metrics Servlets .................. SKIPPED [INFO] Kite Morphlines Saxon ............................. SKIPPED [INFO] Kite Morphlines Tika Core ......................... SKIPPED [INFO] Kite Morphlines Tika Decompress ................... SKIPPED [INFO] Kite Morphlines Twitter ........................... SKIPPED [INFO] Kite Morphlines Hadoop Core ....................... SKIPPED [INFO] Kite Morphlines Hadoop Record Columnar (RC) File .. SKIPPED [INFO] Kite Morphlines Hadoop Sequence File .............. SKIPPED [INFO] Kite Morphlines UserAgent ......................... SKIPPED [INFO] Kite Morphlines Solr Core ......................... SKIPPED [INFO] Kite Morphlines Solr Cell ......................... SKIPPED [INFO] Kite Morphlines Dependency Aggregator ............. SKIPPED [INFO] Kite Morphlines Dependency Aggregator (Excluding Solr) SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 7.192s [INFO] Finished at: Sat Dec 14 21:37:29 CET 2013 [INFO] Final Memory: 33M/346M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal on project kite-data-hbase: Could not resolve dependencies for project org.kitesdk:kite-data-hbase:jar:0.10.1-SNAPSHOT: Failure to find org.kitesdk:kite-data-core:jar:0.10.1-SNAPSHOT in https://repository.cloudera.com/artifactory/cloudera-repos/ was cached in the local repository, resolution will not be reattempted until the update interval of com.cloudera.releases has elapsed or updates are forced -> [Help 1] what am I missing, or doing wrong? thanks...Gerd...
... View more
Labels:
12-12-2013
11:54 AM
Hi Philip, many thanks for your additional explanation, sounds absolutely reasonable. Keep up your great work for CM, brilliant tool best regards...Gerd...
... View more
12-12-2013
01:23 AM
Hi, I'm curious if it is possible to solve this problem with Flume: I have a SpoolingDir source where files with names in the format "prefixA.prefixB.importantPart.csv" will be moved to The files shall be put into HDFS (with its original filename) into the corresponding directory "hdfs://basepath/importantPart/", so that the absolute path for a file is "hdfs://basepath/importantPart/prefixA.prefixB.importantPart.csv". a) how can I parse the filename to extract "importantPart" to create the output HDFS path accordingly, or is this possible at all with Flume? b) how to preserve the original filename so that the HDFS sink writes to the file with the same filename, again, possible at all? Yes, I know, Flume isn't the right tool for such "file copy" approaches it's working on events, but nevertheless it is interesting if it is possible or if someone did this already. Any hint highly appreciated....
... View more
Labels:
- Labels:
-
Apache Flume
-
HDFS
12-12-2013
12:01 AM
Hi clint, thanks for your response. Yes, the link you provided is exactly the "how-to" that I followed, and at the end everything is working fine. What I am wondering about is the fact that the HostInspector isn't determining the used java version from the running processes directly. Just relying on the first JDK found (from /usr/lib64/cmf/service/common/cloudera-config.sh) can definitely lead to wrong results, I mean the cluster is running and working fine on Java7 but the HostInspector tells the nodes are using Java6 just because the new path hasn't been added to this script, isn't that weird ?
... View more
12-11-2013
12:38 PM
1 Kudo
Hi, today I successfully upgraded our cluster to CM4.8.0 and in parallel to Java7 (1.7.0_45, Oracle7). Everything ran smoothly and as expected. Except one thing: after the upgrade I started the HostInspector and it told me that all nodes are using still the old java version 1.6.0_36 HostInspector is just kidding, I thought, but nope....after restarting the whole cluster the HostInspector still reported the old version. But ALL the services on the cluster nodes have been started using java-1.7.0_45. To achive this I set the JAVA_HOME property on each "service environment safety valve" and I verified it by executing "ps auxf | grep java" on the nodes. I was wondering why the HostInspector still reports a different java version, and, more important, how he detects the version. After adding the Java7 path to the file /usr/lib64/cmf/service/common/cloudera-config.sh everything was fine, HostInspector now shows the correct java version for all nodes. Now my question: why does the HostInspector just rely on a setting in the mentioned shell script to determine (and display) the java version the cluster services are running with ? Is this on purpose or is this a bug ?!?! I'd expect that the HostInspector is checking the REAL USED jvm of the running Hadoop service processes and displays it. It's a weak source of truth if it reports a different version than the output of the process list tells. What do you think? many thanks...Gerd...
... View more
Labels:
- Labels:
-
Apache Hadoop
11-15-2013
03:04 AM
Hi, since I didn't find a proper way to detect the version number of the installed components (is there a way to gather this?), I'd suggest to e.g. add a column in the general service overview. What do you think? regards...Gerd...
... View more
11-02-2013
01:27 PM
Hi again, problem is solved, yiieeah ! Additionally to Abe's hint I detected that the sqoop server parameter on the hue.ini entry was again missing. After adding this value back everything's fine and the sqoop wizard launches. Thanks to all who participated in my troubles
... View more