Member since 
    
	
		
		
		04-22-2016
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                931
            
            
                Posts
            
        
                46
            
            
                Kudos Received
            
        
                26
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 1848 | 10-11-2018 01:38 AM | |
| 2209 | 09-26-2018 02:24 AM | |
| 2240 | 06-29-2018 02:35 PM | |
| 2910 | 06-29-2018 02:34 PM | |
| 6088 | 06-20-2018 04:30 PM | 
			
    
	
		
		
		09-22-2018
	
		
		02:50 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 ok I am using binary mode so I will use port 10000. its currently blocked so have to ask security guys to open it.  the hive.server2.authentication is not set but I see the two other similar variables .  so for authentication method I would just choose "username"  and not "username/password" ?    where would I create this user?      we are not using Kerberos.       
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-22-2018
	
		
		04:36 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I read your DATA ACCESS document and I think its for cases where you don't have a date check column.  since its doing many operations like merging tables ,  purging , compacting , deleting  ..etc   why would I do all this when I can just    1- import the whole base table as ORC    2-  bring in the incrementals as text  to an exterenally mapped table   3- insert into the base ORC table selecting everything from the incremental table   4- delete all the files in the external table folder.    I tested this method and its working fine .  is there any flaw in this method that I am not seeing ? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-22-2018
	
		
		01:26 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 please see the syntax in the attached screenshot ..  its not complaining about dash-dash but its not liking the --append-mode with HCatalog        
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-22-2018
	
		
		01:18 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 here is one of the post showing  hive incremental import into ORC is possible using sqoop . so why its not working for me ?  I was using the correct syntax  . no space between dash-dash   https://community.hortonworks.com/questions/58015/sqoop-hcataloghive-incremental-import-in-orc-forma.html  kindly check my other post ,,  I am trying to follow the link you posted earlier but getting errors.  https://community.hortonworks.com/questions/223262/malformed-orc-file-format.html 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-22-2018
	
		
		01:11 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 here is my sqoop command .    sqoop job -Dmapreduce.job.user.classpath.first=true --create incjob2  -- import --connect "jdbc:oracle:thin:@(description=(address=(protocol=tcp)(host=patronQA)(port=1526))(connect_data=(service_name=patron)))" --username PATRON  --incremental append --check-column INSERT_TIME --table PATRON.UFM -split-by UFM.UFMID  --hcatalog-storage-stanza "stored as orcfile" --compression-codec snappy  --target-dir /user/sami
  here is my create external table command   CREATE EXTERNAL TABLE IF NOT EXISTS ufm_orc (
..
..
 )
STORED AS ORC location '/user/sami'
  here is the error , as you can see both table input and output format is ORC   SerDe Library:          org.apache.hadoop.hive.ql.io.orc.OrcSerde
InputFormat:            org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
OutputFormat:           org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
Compressed:             No
Num Buckets:            -1
Bucket Columns:         []
Sort Columns:           []
Storage Desc Params:
        serialization.format    1
Time taken: 0.495 seconds, Fetched: 217 row(s)
    > select ufmid,insert_time from ufm_orc limit 10;
OK
Failed with exception java.io.IOException:org.apache.hadoop.hive.ql.io.FileFormatException: Malformed ORC file hdfs://hadoop1.tolls.dot.state.fl.us:8020/user/sami/part-m-00000.snappy. Invalid postscript.
Time taken: 0.328 seconds
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
	
					
			
		
	
	
	
	
				
		
	
	
			
    
	
		
		
		09-21-2018
	
		
		09:25 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 it must be my lucky day 🙂  Stanca but this will introduce a lot of delays in the data , we wanted a near real time data , is it not possible using sqoop ?  what about Nifi ?  Also i use "--hcatalog-storage-stanza 'stored as orc tblproperties ("orc.compress"="SNAPPY")'"  for non incremental loads and i was told that soon this would be working  , still not ? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		09-21-2018
	
		
		09:15 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 i dont have SASL enabled so i am using http to connect , can you please take a look at the screenshots and see what i am doing wrong ?  1- in the port what should i use?  10000 , 10001 or 9083 ?  2-  the username and password is the ambari login username/password or an hive username?  also which of the ports should be open between client and hive server ?             
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
	
					
			
		
	
	
	
	
				
		
	
	
			
    
	
		
		
		09-21-2018
	
		
		08:17 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I have a strong feeling Hortonworks is purposely not answering it  as I got all my other questions answered ..so I m hoping one brave person will step forward and tell me the truth . after all that's what the purpose of the forum is.  my question that is still not answered is :  is sqoop incremental load in hive ORC table supported and has anyone done it ?   i am sure many people will benefit from this answer  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
	
					
			
		
	
	
	
	
				
		
	
	
			
    
	
		
		
		06-29-2018
	
		
		06:11 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 oh the following syntax worked   [root@hadoop1 ~]# curl --negotiate -i -u : -X GET -H "Accept: text" http://$(hostname):17001/
HTTP/1.1 401 Authentication required
WWW-Authenticate: Negotiate
Set-Cookie: hadoop.auth=; Path=/; HttpOnly
Content-Type: text/html; charset=iso-8859-1
Cache-Control: must-revalidate,no-cache,no-store
Content-Length: 1393
HTTP/1.1 200 OK
Set-Cookie: hadoop.auth="u=hbase&p=hbase/hadoop1.xxx.com@XXX.US&t=kerberos&e=1530331783162&s=Ypuvww45JSzCbQwTbc5ysWmaSfI="; Path=/; HttpOnly
Content-Type: text/plain
Cache-Control: no-cache
Content-Length: 18
UFM
WZ
state_code
   
						
					
					... View more