Member since 
    
	
		
		
		05-07-2018
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                331
            
            
                Posts
            
        
                45
            
            
                Kudos Received
            
        
                35
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 9613 | 09-12-2018 10:09 PM | |
| 3742 | 09-10-2018 02:07 PM | |
| 11507 | 09-08-2018 05:47 AM | |
| 4090 | 09-08-2018 12:05 AM | |
| 4932 | 08-15-2018 10:44 PM | 
			
    
	
		
		
		06-07-2018
	
		
		03:00 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Awesome! 🙂 Good to know @Souveek Ray 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-07-2018
	
		
		02:45 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Good to know @Rahul Kumar!   I already used in cluster mode and it's quite good as well.  So keep going on this! It's an interesting feature from kafka  PS: I'm not an expert yet haha, just a humble guy trying to help xD  Let us know if you have more questions! 🙂  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		11:12 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hey @cskbhatt  Did you tried to execute the following line?   ./bin/install-interpreter.sh --name jdbc --artifact org.apache.zeppelin:zeppelin-jdbc:0.7.3   Here's the links that i've searched.  https://mvnrepository.com/artifact/org.apache.zeppelin/zeppelin-jdbc/0.7.3  https://zeppelin.apache.org/docs/0.7.3/manual/interpreterinstallation.html  Hope this helps! 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		10:52 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hey @Paramesh malla !  Could check if your yarn.resourcemanager.recovery.enabled is true? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		10:41 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hey @Jason Sphar !   Hmm that's strange, did you try to run with zookeeper parameter?   /bin/kafka-consumer-groups.sh --zookeeper ZKHOSTS:2181 --describe --group console-consumer-96628
  Hope this helps! 🙂 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		07:15 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Jason Sphar   And could you check if there's any extra consumer group (probably used by Nifi Processor)   Sorry about the mess, actually I mean by trying to see in command line if there's any consumer group for nifi processor.   You can try this one, to see if Nifi is able to reach kafka and create a consumer group.  /usr/<yourKafkapath>/kafka/bin/kafka-consumer-groups.sh --bootstrap-server kafka-host:9092 --list  And if you found smtg, can try to describe it:  /usr/<yourKafkapath>/kafka/bin/kafka-consumer-groups.sh --bootstrap-server kafka-hotst:9092 --describe --group consumer_anyname_2390934  Kafka is running on a Free Tier Linux within in Amazon Ec2 so I know there are restrictions of size and memory usages but since I was able to configure Kafka to use less memory it works fine within that local environment. I don't think that is the issue here but just thought to mention it.  Yeah agree on that, It shouldn't be a problem, since you can work with kafka manually.  Hope this helps 🙂 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		06:27 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi @Thierry Vernhet  Do you have more examples from your HDFS records? 🙂     I made a simple test here, and its seems to be working your REGEX with the Uppercase  Here's the output  CREATE EXTERNAL TABLE my_arming_table 
( dc_syslog_date STRING, 
dc_syslog_hostname STRING, 
dc_syslog_process STRING, 
dc_logtype STRING, 
dc_message STRING) 
PARTITIONED BY (yearbrut INT, monthbrut INT, daybrut INT)
ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.RegexSerDe'
WITH SERDEPROPERTIES( 'input.regex'='^(\\S+)\\s(\\S+)\\s(\\S+)\\s(ARMING)\\s(.*)') STORED AS TEXTFILE
LOCATION '/user/hive/warehouse/my_arming_table/';
ALTER TABLE my_arming_table ADD PARTITION(yearbrut=2018,monthbrut=06,daybrut=06);
#Command to put some data in the table hdfs dfs -put regex_hcc  /user/hive/warehouse/my_arming_table/yearbrut=2018/monthbrut=6/daybrut=6/
#And here's a quick describe from my tablehive> desc formatted my_arming_table;
OK
# col_name            	data_type           	comment             
	 	 
dc_syslog_date      	string              	                    
dc_syslog_hostname  	string              	                    
dc_syslog_process   	string              	                    
dc_logtype          	string              	                    
dc_message          	string              	                    
	 	 
# Partition Information	 	 
# col_name            	data_type           	comment             
	 	 
yearbrut            	int                 	                    
monthbrut           	int                 	                    
daybrut             	int                 	                    
	 	 
# Detailed Table Information	 	 
Database:           	default             	 
Owner:              	hive                	 
CreateTime:         	Wed Jun 06 18:00:02 UTC 2018	 
LastAccessTime:     	UNKNOWN             	 
Protect Mode:       	None                	 
Retention:          	0                   	 
Location:           	hdfs://vini-horton/user/hive/warehouse/my_arming_table	 
Table Type:         	EXTERNAL_TABLE      	 
Table Parameters:	 	 
	EXTERNAL            	TRUE                
	transient_lastDdlTime	1528308002          
	 	 
# Storage Information	 	 
SerDe Library:      	org.apache.hadoop.hive.serde2.RegexSerDe	 
InputFormat:        	org.apache.hadoop.mapred.TextInputFormat	 
OutputFormat:       	org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat	 
Compressed:         	No                  	 
Num Buckets:        	-1                  	 
Bucket Columns:     	[]                  	 
Sort Columns:       	[]                  	 
Storage Desc Params:	 	 
	input.regex         	^(\\S+)\\s(\\S+)\\s(\\S+)\\s(ARMING)\\s(.*)
	serialization.format	1   
 
#Finally the select outputhive>select * from my_arming_table;
OK
2018-06-06T11:28:54+02:00	sazqye	named[980]:	ARMING	trigger on (action:LOG) (T6-Recursive-attacks recursive-time: 1283)	2018	6	6
NULL	NULL	NULL	NULL	NULL	2018	6	6
Time taken: 0.175 seconds, Fetched: 2 row(s)	
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		05:48 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Hey @Rahul Kumar could you try to run as kafka user?  Or root  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		05:43 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hey @Jason Sphar what about the port from zookeeper? (2181 by default).   And could you check if there's any extra consumer group (probably used by Nifi Processor).   Hope this helps!  Ah and kafka logs are showing something?  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-06-2018
	
		
		05:22 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hey @Souveek Ray  Maybe this code can help you (it's just a scratch)  select 
concat_ws(' ',case when regexp_extract('abcd pqrs xyzsd mnop', '(abcd)', 0) is null then '' else 'abcd' end, 
case when regexp_extract('abcd pqrs xyzsd mnop', '(mnop)', 0) is null then '' else 'mnop' end,
case when regexp_extract('abcd pqrs xyzsd mnop', '(pqrs)', 0) is null then '' else 'pqrs' end)
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		- « Previous
- Next »
 
         
					
				













