Member since
11-24-2015
223
Posts
10
Kudos Received
0
Solutions
03-06-2019
07:10 PM
>this file should be created locally on the node manager where your shell script was run I have the file locally on a node. It is on this node that I execute the oozie command. Are you saying that the shell script should be in hdfs? And that yarn will execute this shell script on the node manager on which this job was run? Rather than on the host from which I run the oozie command? Appreciate the feedback.
... View more
03-05-2019
06:18 PM
In the yarn log for the job I can see : ================================================================= >>> Invoking Shell command line now >> Exit code of the Shell command 0 <<< Invocation of Shell command completed <<< <<< Invocation of Main class completed <<< Oozie Launcher, capturing output data: =======================
... View more
03-05-2019
02:54 PM
I created a similar job as in : https://community.hortonworks.com/content/supportkb/48985/how-to-setup-oozie-shell-action.html On execution the yarn log shows as success. But I can't see the /tmp/output file being created anywhere. I checked on the local linux host as well as hdfs. One question I have is in job.properties - oozie.coord.application.path=${nameNode}/user/root/apps/shell Should the configuration files (job.properties, coordinator.xml etc) reside in the above directory? I have them there. Not sure what is happening. Appreciate the insights.
... View more
Labels:
- Labels:
-
Apache Oozie
12-19-2018
06:28 PM
@Naveen my hbase table is created only with a single column family. create 'mbev:hb_test' , 'me_data' Then data is fed into the hbase table through flume. This data originally was sent to a hive table with 10 columns - but for small files issue, i am redirecting the same data to the above hbase table. my flume config has the following lines : r_hbase.sinks.sink1.table = mbev:hb_test r_hbase.sinks.sink1.columnFamily = me_data r_hbase.sinks.sink1.serializer.columns =col1,col2,col3,col4,col5,col6,col7,col8,col9,col10 So I can see data coming into the hbase table. Data is showing like : hbase(main):008:0> scan "mbev:hb_test" ROW COLUMN+CELL default0998c6b9-2fa9 column=me_data:pCol, timestamp=1545242255268, value=.z3knt
-122e-1536-0o42ef7fb dErc90GqYg5a3n-zTQ\a04NQA32018-12-19-12.57.30.000123\x09FY
a3e YY\q1V10006002079317\x01M\x00\x09KEF\x00\x00 So how do I create the hive table to see the data in the hbase table. I tried as we did above initially with only "":key,me_data:id", but don't see the data in the hive table. I also tried : create external table tmp_test4 (col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,me_data:col1,me_data:col2,me_data:col3,me_data:col4,me_data:col5,me_data:col6,me_data:col7,me_data:col8,me_data:col9,me_data:col10")
TBLPROPERTIES("hbase.table.name" = "mbev:hb_test"); FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. java.lang.RuntimeException:
MetaException(message:org.apache.hadoop.hive.serde2.SerDeException org.apache.hadoop.hive.hbase.HBaseSerDe: columns has 10 elements while hbase.columns.
mapping has 11 elements (counting the key if implicit)) Appreciate your help.
... View more
12-18-2018
08:49 PM
btw : the create table shows two columns : create external table hv_test (id string, idate string) so how can this statement show only 1 column : WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,me_data:id") don't we need to likewise map idate?
... View more
12-18-2018
04:51 PM
In my hbase table actually data for 10 columns lands from flume. In the flume configuration I have defined : host1.sinks.sink1.serializer.columns = col1, col2 ... In HBASE the table is only defined as above : create 'mbev:hb_test' , 'me_data' So how do I define the hive table for the respective data for the 10 columns?
... View more
12-17-2018
06:07 PM
I created a table in hbase. Created a table in hive mapping it to the table in hbase. Inserted a row in the hbase table. But hive table doesn't show any rows. Any idea what I am doing wrong? Appreciate the feedback. hbase(main):019:0> create 'mbev:hb_test' , 'me_data' 0 row(s) in 1.2670 seconds hbase(main):006:0> put "mbev:hb_test",'1',"me_data:id",'1' 0 row(s) in 0.1700 seconds hbase(main):007:0> scan "mbev:hb_test" ROW COLUMN+CELL
1 column=me_data:id, timestamp=1545064141017, value=1
1 row(s) in 0.0590 seconds hive>create external table hv_test (id string, idate string)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,me_data:idate") TBLPROPERTIES("hbase.table.name" = "mbev:hb_test"); hive> select * from hv_test ; OK
Time taken: 0.23 seconds
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Hive