Member since
07-21-2014
141
Posts
8
Kudos Received
3
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2771 | 02-01-2017 04:49 PM | |
2080 | 01-15-2015 01:57 PM | |
2729 | 01-05-2015 12:59 PM |
01-15-2015
12:52 PM
I'm using the patch from FLUME-2578 but when I pass the patched jar via --classpath option of flume-ng, I see the old jar being used instead of the patched jar since it is being appended. Is there a way to override or append the patched jar to java classpath and run flume agent? Thanks!
... View more
Labels:
- Labels:
-
Apache Flume
-
Apache Kafka
01-12-2015
10:34 PM
In Cloudera Manager, where can I find the list of available versions of services (Flume, Hive, Impala etc.) are installed in the cluster? Thanks!
... View more
Labels:
01-05-2015
12:59 PM
Found the metrics under Cloudera Manager->Flume->Charts Library Thanks!
... View more
01-05-2015
10:07 AM
I'm currently using Flume with CDH 5.3.0 with Kite Dataset sink to store as Avro. I would like to get some insights into the data flowing into the cluster. How do get any sort of metrics on the data throughput, performance etc ? Thanks!
... View more
Labels:
- Labels:
-
Apache Flume
12-11-2014
04:10 PM
Does Impala 2.0 support nested data like map or is it expected in future releases? Thanks!
... View more
Labels:
- Labels:
-
Apache Impala
08-04-2014
12:16 PM
Thanks Sean.. I'm currently computing uniques visitors per page and running a count distinct using SparkSQL. We also run the non-spark jobs on the cluster, so if we allocate the 2GB I'm assuming we can't run any other jobs simultaneously. Also, I'm also looking to see how to set the storage levels in CM.
... View more
08-04-2014
10:38 AM
I got a 40 node cdh 5.1 cluster and attempting to run a simple spark app that processes about 10-15GB raw data but I keep running into this error: java.lang.OutOfMemoryError: GC overhead limit exceeded Each node has 8 cores and 2GB memory. I notice the heap size on the executors is set to 512MB with total set to 2GB. Wanted to know whats the heap size needs to be set to for such data sizes? Thanks for the input!
... View more
Labels:
07-28-2014
10:43 AM
After removing the import, I was able to compile the package successfully.
... View more
07-28-2014
10:38 AM
Thanks Sean, now I get this: error: object SQLContext is not a member of package org.apache.spark.sql [INFO] Note: class SQLContext exists, but it has no companion object. [INFO] import org.apache.spark.sql.SQLContext._
... View more
07-28-2014
10:28 AM
I'm creating a simple SparkSQL app based on this post by Sandy: http://blog.cloudera.com/blog/2014/04/how-to-run-a-simple-apache-spark-app-in-cdh-5/ But 'mvn package' gives throws error: error: object sql is not a member of package org.apache.spark Any idea if I need to include any other dependency? Thanks!
... View more
Labels:
- « Previous
- Next »