Member since 
    
	
		
		
		09-01-2016
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                44
            
            
                Posts
            
        
                3
            
            
                Kudos Received
            
        
                0
            
            
                Solutions
            
        
			
    
	
		
		
		07-28-2018
	
		
		02:32 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Kafka has 10k messages left over due to Kafka consumer (topology) is down.   I dont want to process the left over messages when topology is up. How to remove those messages.  One more use case, I want to show demo to user, I want to quickly delete those messages from kafka.   how to delete kafka messages instantly. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Kafka
			
    
	
		
		
		09-19-2017
	
		
		03:28 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Bala Vignesh N V , Thanks.      In case, I am running same hql queries daily, then I need to store the result in the file . once hive tasks are completed then read the file and update the table ?  and any other way to achieve it. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-19-2017
	
		
		02:16 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi @Bala Vignesh N V ,       looking for detailed info, assume below are the query in hql ,            select  count(*) from table1;     
       select sum(col1) from table1 group by col2 ;   
       select sum(col3) from table1 group by col4          first query will returns single column which has count ,  remaining returns 10  columns each.  where this result will store and how to process and more importantly what is the recommended way handle this or how industry handling this ?  Thanks.             
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-18-2017
	
		
		03:34 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hello Geeks,     Need clarifications on below question        1. is hql used to perform multiple quires in single go ?       2. if yes, How do i set hint for different quires ?  ex: first query I have to set, SMB join , on second I have to set skew join. ( confused about , the property I have set  for first query is also applicable for second query )       3. all the 10 quires are returned results , how to process the result. how industry handling this ?      
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Hive
			
    
	
		
		
		09-04-2017
	
		
		05:48 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hi All,      I have given 8GB ram, but still it takes so much of time to load. pls find the attached screen shot.   VM : Vmware   HDP : hdp2.6     
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-05-2017
	
		
		01:14 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							@Namit Maheshwari Yes, there is a pattern for creating partition (yyyy-mm-dd) ..   OK,your idea is , run the command and store the result and check for the existence of the partition ??  Is there any other simple way to check ?  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-05-2017
	
		
		03:30 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Thanks for your answer, I am looking for kind of API to tell that, partition added or not.. I will be difficult to check the partition status daily and take call when there is a failure. ( Want to automate rather checking it manually) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-05-2017
	
		
		03:11 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Created table in Hive with dynamic partition enabled..  Adding partition on daily basis  ALTER TABLE test ADD PARTITION (date='2014-03-17') location 's3://test.com/2014-03-05'  Now ,How do I ensure that whether partition added or not ?  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Hive
			
    
	
		
		
		11-22-2016
	
		
		03:16 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Vedant Jain , @Bernhard Walter       well, may be my question is misleading you , let me elaborate it.  val textFile = sc.textFile("hdfs://...") 
val counts = textFile.flatMap(line => line.split(" "))
                 .map(word => (word, 1))
                 .reduceByKey(_ + _) 
counts.saveAsTextFile("hdfs://...")  a simple wordcount problem.. this piece of code given to driver program, which creates DAG and stages, and given task to respective worker nodes where actual operation is happening.  Now, lets look at the first line of the program. From the file, RDD is generated (SparkContext implemented textFile() function  which generates RDD from file).  file is resides in worker node. from worker node, we needs to get the RDD out.  In order to acheive that , Worker node ( or executor ) needs to have the SparkContext, Isn't it ?  My Question is, How does executor gets the spark context ? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		11-21-2016
	
		
		02:33 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi All ,     I am creating spark context in the driver, How does executor gets the spark context ?    Can anyone share any link with this context will be helpful to understand the system much better. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Spark
 
        









