Member since
09-30-2016
11
Posts
1
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
4700 | 11-14-2016 02:49 PM | |
1300 | 11-03-2016 07:29 PM |
11-14-2016
02:49 PM
I finally figured this out and thought it would be friendly of me to post the solution. One of those that when you finally get it you think, "Ugh, that was so obvious". One important note, if you are having trouble with Hive make sure to check the Yarn logs too! My solution to this and so many other issues was ensuring all my nodes had all the other nodes ip addresses in their host files. This ensures Ambari picks up all the correct IPs by hostname. I am on Ubuntu so I did the following: $ vim /etc/hosts And then the file came out looking like this: 127.0.0.1 localhost
#127.0.1.1 ambarihost.com ambarihost
# Assigning static IP here so ambari gets it right
192.168.0.20 ambarihost.com ambarihost
#Other hadoop nodes
192.168.0.21 kafkahost.com kafkahost
192.168.0.22 hdfshost.com hdfshost
... View more
11-11-2016
06:14 PM
No, there isn't. That would have been nice, wouldn't it?
... View more
11-11-2016
02:49 PM
I had permissions issues I had created here because the original issue is back. It is this one: org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-300459168-127.0.1.1-1478287363661:blk_1073741947_1125 file=/tmp/hive/hiveuser/_tez_session_dir/34896acf-3209-4aa0-a244-5a28b5b15b92/hive-hcatalog-core.jar So that is my real question, I guess!
... View more
11-11-2016
01:47 PM
I am running an insert into table select from... query on Hive. Whether I set the execution engine to TEZ or MR I am getting BlockMissingException errors. They all look similar to this one: Diagnostics: org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-300459168-127.0.1.1-1478287363661:blk_1073741827_1003 file=/hdp/apps/2.5.0.0-1245/tez/tez.tar.gz When I go into HDFS the files are there. They exist. So I thought maybe it is a permissions issue. But all my related proxyusers are set to hosts=* and groups=* just to try to rule this all out. I have a 2.5 cluster hosted on Ubuntu 12.04. Can anyone point me in a direction of what I might be missing here?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
-
Apache Tez
11-03-2016
07:29 PM
Okay, I found a workaround. Since I don't really need this table to be transactional, it was just a nice to have, I have created the table without buckets and without TBLPROPERTIES and now it works as expected. create table stgicplogs (actdatetime timestamp,server VARCHAR(10),pid VARCHAR(25),level VARCHAR(50),type VARCHAR(50),details VARCHAR(8000)) PARTITIONED BY(actdate DATE) STORED AS orc;
... View more
11-03-2016
06:15 PM
I am doing a multistep loading process to get data into a transactional table. I have the two table defintions shown at the bottom of this questio. When I execute "INSERT INTO TABLE stgicplogs PARTITION (actdate='2016-03-28') select * from stgicplogs_stg;" it appears to go well. There are no errors. However, there is also no data in stgicplogs. Data does exist in stgicplogs_stg and returns when selected nicely. Any idea what I might be missing? I do have hive.exec.dynamic.partition.mode set to nonstrict and hive.enforce.bucketing set to true. The result after running the insert statement in beeline is: INFO : Session is already open
INFO : Dag name: INSERT INTO TABLE stgicplog...stgicplogs_stg(Stage-1)
INFO :
INFO : Status: Running (Executing on YARN cluster with App id application_1478118527771_0011)
INFO : Map 1: 0/1 Reducer 2: 0/20
INFO : Map 1: 0/1 Reducer 2: 0/20
INFO : Map 1: 0(+1)/1 Reducer 2: 0/20
INFO : Map 1: 0(+1)/1 Reducer 2: 0/20
INFO : Map 1: 1/1 Reducer 2: 0/20
INFO : Map 1: 1/1 Reducer 2: 0(+1)/20
INFO : Map 1: 1/1 Reducer 2: 1(+0)/20
INFO : Map 1: 1/1 Reducer 2: 1(+1)/20
INFO : Map 1: 1/1 Reducer 2: 2(+0)/20
INFO : Map 1: 1/1 Reducer 2: 2(+1)/20
INFO : Map 1: 1/1 Reducer 2: 3(+1)/20
INFO : Map 1: 1/1 Reducer 2: 4(+0)/20
INFO : Map 1: 1/1 Reducer 2: 4(+1)/20
INFO : Map 1: 1/1 Reducer 2: 5(+1)/20
INFO : Map 1: 1/1 Reducer 2: 6(+1)/20
INFO : Map 1: 1/1 Reducer 2: 7(+0)/20
INFO : Map 1: 1/1 Reducer 2: 7(+1)/20
INFO : Map 1: 1/1 Reducer 2: 8(+0)/20
INFO : Map 1: 1/1 Reducer 2: 8(+1)/20
INFO : Map 1: 1/1 Reducer 2: 9(+0)/20
INFO : Map 1: 1/1 Reducer 2: 9(+1)/20
INFO : Map 1: 1/1 Reducer 2: 9(+2)/20
INFO : Map 1: 1/1 Reducer 2: 10(+2)/20
INFO : Map 1: 1/1 Reducer 2: 11(+2)/20
INFO : Map 1: 1/1 Reducer 2: 12(+1)/20
INFO : Map 1: 1/1 Reducer 2: 12(+2)/20
INFO : Map 1: 1/1 Reducer 2: 12(+3)/20
INFO : Map 1: 1/1 Reducer 2: 13(+3)/20
INFO : Map 1: 1/1 Reducer 2: 14(+3)/20
INFO : Map 1: 1/1 Reducer 2: 14(+5)/20
INFO : Map 1: 1/1 Reducer 2: 15(+4)/20
INFO : Map 1: 1/1 Reducer 2: 15(+5)/20
INFO : Map 1: 1/1 Reducer 2: 16(+4)/20
INFO : Map 1: 1/1 Reducer 2: 17(+3)/20
INFO : Map 1: 1/1 Reducer 2: 18(+2)/20
INFO : Map 1: 1/1 Reducer 2: 19(+1)/20
INFO : Map 1: 1/1 Reducer 2: 20/20
INFO : Loading data to table default.stgicplogs partition (actdate=2016-03-28) from hdfs://localhost:8020/apps/hive/warehouse/stgicplogs/actdate=2016-03-28/.hive-staging_hive_2016-11-03_13-54-35_964_9082514218818653049-1/-ext-10000
INFO : Partition default.stgicplogs{actdate=2016-03-28} stats: [numFiles=5, numRows=0, totalSize=8626, rawDataSize=0]
No rows affected (31.42 seconds)
create table stgicplogs (actdatetime timestamp,server VARCHAR(10),pid VARCHAR(25),level VARCHAR(50),type VARCHAR(50),details VARCHAR(8000)) PARTITIONED BY(actdate DATE) CLUSTERED BY(server) INTO 20 BUCKETS STORED AS orc TBLPROPERTIES("transactional"="true");
create table stgicplogs_stg (actdatetime timestamp,server VARCHAR(10),pid VARCHAR(25),level VARCHAR(50),type VARCHAR(50),details VARCHAR(8000), actdate DATE) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.OpenCSVSerde' WITH SERDEPROPERTIES ("separatorChar" = "|","quoteChar" = "'","escapeChar" = "\\");
... View more
Labels:
- Labels:
-
Apache Hive