Member since
09-01-2016
44
Posts
3
Kudos Received
0
Solutions
07-28-2018
02:32 AM
Kafka has 10k messages left over due to Kafka consumer (topology) is down. I dont want to process the left over messages when topology is up. How to remove those messages. One more use case, I want to show demo to user, I want to quickly delete those messages from kafka. how to delete kafka messages instantly.
... View more
Labels:
- Labels:
-
Apache Kafka
09-19-2017
03:28 PM
@Bala Vignesh N V , Thanks. In case, I am running same hql queries daily, then I need to store the result in the file . once hive tasks are completed then read the file and update the table ? and any other way to achieve it.
... View more
09-19-2017
02:16 AM
Hi @Bala Vignesh N V , looking for detailed info, assume below are the query in hql , select count(*) from table1;
select sum(col1) from table1 group by col2 ;
select sum(col3) from table1 group by col4 first query will returns single column which has count , remaining returns 10 columns each. where this result will store and how to process and more importantly what is the recommended way handle this or how industry handling this ? Thanks.
... View more
09-18-2017
03:34 AM
Hello Geeks, Need clarifications on below question 1. is hql used to perform multiple quires in single go ? 2. if yes, How do i set hint for different quires ? ex: first query I have to set, SMB join , on second I have to set skew join. ( confused about , the property I have set for first query is also applicable for second query ) 3. all the 10 quires are returned results , how to process the result. how industry handling this ?
... View more
Labels:
- Labels:
-
Apache Hive
09-04-2017
05:48 AM
1 Kudo
Hi All, I have given 8GB ram, but still it takes so much of time to load. pls find the attached screen shot. VM : Vmware HDP : hdp2.6
... View more
04-05-2017
01:14 PM
@Namit Maheshwari Yes, there is a pattern for creating partition (yyyy-mm-dd) .. OK,your idea is , run the command and store the result and check for the existence of the partition ?? Is there any other simple way to check ?
... View more
04-05-2017
03:30 AM
Thanks for your answer, I am looking for kind of API to tell that, partition added or not.. I will be difficult to check the partition status daily and take call when there is a failure. ( Want to automate rather checking it manually)
... View more
04-05-2017
03:11 AM
Created table in Hive with dynamic partition enabled.. Adding partition on daily basis ALTER TABLE test ADD PARTITION (date='2014-03-17') location 's3://test.com/2014-03-05' Now ,How do I ensure that whether partition added or not ?
... View more
Labels:
- Labels:
-
Apache Hive
11-22-2016
03:16 AM
@Vedant Jain , @Bernhard Walter well, may be my question is misleading you , let me elaborate it. val textFile = sc.textFile("hdfs://...")
val counts = textFile.flatMap(line => line.split(" "))
.map(word => (word, 1))
.reduceByKey(_ + _)
counts.saveAsTextFile("hdfs://...") a simple wordcount problem.. this piece of code given to driver program, which creates DAG and stages, and given task to respective worker nodes where actual operation is happening. Now, lets look at the first line of the program. From the file, RDD is generated (SparkContext implemented textFile() function which generates RDD from file). file is resides in worker node. from worker node, we needs to get the RDD out. In order to acheive that , Worker node ( or executor ) needs to have the SparkContext, Isn't it ? My Question is, How does executor gets the spark context ?
... View more
11-21-2016
02:33 PM
Hi All , I am creating spark context in the driver, How does executor gets the spark context ? Can anyone share any link with this context will be helpful to understand the system much better.
... View more
Labels:
- Labels:
-
Apache Spark