Member since
04-22-2016
931
Posts
46
Kudos Received
26
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1537 | 10-11-2018 01:38 AM | |
1908 | 09-26-2018 02:24 AM | |
1867 | 06-29-2018 02:35 PM | |
2478 | 06-29-2018 02:34 PM | |
5467 | 06-20-2018 04:30 PM |
09-22-2018
02:50 PM
ok I am using binary mode so I will use port 10000. its currently blocked so have to ask security guys to open it. the hive.server2.authentication is not set but I see the two other similar variables . so for authentication method I would just choose "username" and not "username/password" ? where would I create this user? we are not using Kerberos.
... View more
09-22-2018
04:36 AM
I read your DATA ACCESS document and I think its for cases where you don't have a date check column. since its doing many operations like merging tables , purging , compacting , deleting ..etc why would I do all this when I can just 1- import the whole base table as ORC 2- bring in the incrementals as text to an exterenally mapped table 3- insert into the base ORC table selecting everything from the incremental table 4- delete all the files in the external table folder. I tested this method and its working fine . is there any flaw in this method that I am not seeing ?
... View more
09-22-2018
01:26 AM
please see the syntax in the attached screenshot .. its not complaining about dash-dash but its not liking the --append-mode with HCatalog
... View more
09-22-2018
01:18 AM
here is one of the post showing hive incremental import into ORC is possible using sqoop . so why its not working for me ? I was using the correct syntax . no space between dash-dash https://community.hortonworks.com/questions/58015/sqoop-hcataloghive-incremental-import-in-orc-forma.html kindly check my other post ,, I am trying to follow the link you posted earlier but getting errors. https://community.hortonworks.com/questions/223262/malformed-orc-file-format.html
... View more
09-22-2018
01:11 AM
here is my sqoop command . sqoop job -Dmapreduce.job.user.classpath.first=true --create incjob2 -- import --connect "jdbc:oracle:thin:@(description=(address=(protocol=tcp)(host=patronQA)(port=1526))(connect_data=(service_name=patron)))" --username PATRON --incremental append --check-column INSERT_TIME --table PATRON.UFM -split-by UFM.UFMID --hcatalog-storage-stanza "stored as orcfile" --compression-codec snappy --target-dir /user/sami
here is my create external table command CREATE EXTERNAL TABLE IF NOT EXISTS ufm_orc (
..
..
)
STORED AS ORC location '/user/sami'
here is the error , as you can see both table input and output format is ORC SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde
InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
Compressed: No
Num Buckets: -1
Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
Time taken: 0.495 seconds, Fetched: 217 row(s)
> select ufmid,insert_time from ufm_orc limit 10;
OK
Failed with exception java.io.IOException:org.apache.hadoop.hive.ql.io.FileFormatException: Malformed ORC file hdfs://hadoop1.tolls.dot.state.fl.us:8020/user/sami/part-m-00000.snappy. Invalid postscript.
Time taken: 0.328 seconds
... View more
Labels:
09-21-2018
09:25 PM
it must be my lucky day 🙂 Stanca but this will introduce a lot of delays in the data , we wanted a near real time data , is it not possible using sqoop ? what about Nifi ? Also i use "--hcatalog-storage-stanza 'stored as orc tblproperties ("orc.compress"="SNAPPY")'" for non incremental loads and i was told that soon this would be working , still not ?
... View more
09-21-2018
09:15 PM
i dont have SASL enabled so i am using http to connect , can you please take a look at the screenshots and see what i am doing wrong ? 1- in the port what should i use? 10000 , 10001 or 9083 ? 2- the username and password is the ambari login username/password or an hive username? also which of the ports should be open between client and hive server ?
... View more
Labels:
09-21-2018
08:17 PM
I have a strong feeling Hortonworks is purposely not answering it as I got all my other questions answered ..so I m hoping one brave person will step forward and tell me the truth . after all that's what the purpose of the forum is. my question that is still not answered is : is sqoop incremental load in hive ORC table supported and has anyone done it ? i am sure many people will benefit from this answer
... View more
Labels:
06-29-2018
06:11 PM
oh the following syntax worked [root@hadoop1 ~]# curl --negotiate -i -u : -X GET -H "Accept: text" http://$(hostname):17001/
HTTP/1.1 401 Authentication required
WWW-Authenticate: Negotiate
Set-Cookie: hadoop.auth=; Path=/; HttpOnly
Content-Type: text/html; charset=iso-8859-1
Cache-Control: must-revalidate,no-cache,no-store
Content-Length: 1393
HTTP/1.1 200 OK
Set-Cookie: hadoop.auth="u=hbase&p=hbase/hadoop1.xxx.com@XXX.US&t=kerberos&e=1530331783162&s=Ypuvww45JSzCbQwTbc5ysWmaSfI="; Path=/; HttpOnly
Content-Type: text/plain
Cache-Control: no-cache
Content-Length: 18
UFM
WZ
state_code
... View more