Created 02-15-2016 12:04 PM
Hi:
After do this :
egister /usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core-1.2.1.2.3.2.0-2950.jar; register /usr/lib/piggybank/hcatalog-pig-adapter-0.11.0.jar; register /usr/lib/piggybank/hcatalog-core-0.11.0.jar; register /usr/lib/piggybank/hive-shims-0.11.0.jar; . . . F = LOAD 'journey_pig' USING org.apache.hive.hcatalog.pig.HCatLoader();
***The hive-shims-0.11.0.jar contain this classs like this:
public abstract UserGroupInformation getUGIForConf(Configuration paramConfiguration) throws LoginException, IOException;
i receive this error:
2016-02-15 12:54:42,359 [main] ERROR org.apache.pig.tools.grunt.Grunt - ERROR 2998: Unhandled internal error. org.apache.hadoop.hive.shims.HadoopShims.getUGIForConf(Lorg/apache/hadoop/conf/Configuration;)Lorg/apache/hadoop/security/UserGroupInformation; 2016-02-15 12:54:42,359 [main] ERROR org.apache.pig.tools.grunt.Grunt - java.lang.NoSuchMethodError: org.apache.hadoop.hive.shims.HadoopShims.getUGIForConf(Lorg/apache/hadoop/conf/Configuration;)Lorg/apache/hadoop/security/UserGroupInformation; at org.apache.hcatalog.common.HiveClientCache$HiveClientCacheKey.<init>(HiveClientCache.java:201) at org.apache.hcatalog.common.HiveClientCache$HiveClientCacheKey.fromHiveConf(HiveClientCache.java:207) at org.apache.hcatalog.common.HiveClientCache.get(HiveClientCache.java:138) at org.apache.hcatalog.common.HCatUtil.getHiveClient(HCatUtil.java:544) at org.apache.hcatalog.pig.PigHCatUtil.getHiveMetaClient(PigHCatUtil.java:147) at org.apache.hcatalog.pig.PigHCatUtil.getTable(PigHCatUtil.java:183) at org.apache.hcatalog.pig.HCatLoader.getSchema(HCatLoader.java:193) at org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:175) at org.apache.pig.newplan.logical.relational.LOLoad.<init>(LOLoad.java:89) at org.apache.pig.parser.LogicalPlanBuilder.buildLoadOp(LogicalPlanBuilder.java:901) at org.apache.pig.parser.LogicalPlanGenerator.load_clause(LogicalPlanGenerator.java:3568) at org.apache.pig.parser.LogicalPlanGenerator.op_clause(LogicalPlanGenerator.java:1625) at org.apache.pig.parser.LogicalPlanGenerator.general_statement(LogicalPlanGenerator.java:1102) at org.apache.pig.parser.LogicalPlanGenerator.statement(LogicalPlanGenerator.java:560) at org.apache.pig.parser.LogicalPlanGenerator.query(LogicalPlanGenerator.java:421) at org.apache.pig.parser.QueryParserDriver.parse(QueryParserDriver.java:191) at org.apache.pig.PigServer$Graph.parseQuery(PigServer.java:1735) at org.apache.pig.PigServer$Graph.access$000(PigServer.java:1443) at org.apache.pig.PigServer.parseAndBuild(PigServer.java:387) at org.apache.pig.PigServer.executeBatch(PigServer.java:412) at org.apache.pig.PigServer.executeBatch(PigServer.java:398) at org.apache.pig.tools.grunt.GruntParser.executeBatch(GruntParser.java:171) at org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:749) at org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:376) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:230) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:205) at org.apache.pig.tools.grunt.Grunt.exec(Grunt.java:81) at org.apache.pig.Main.run(Main.java:502) at org.apache.pig.Main.main(Main.java:177) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Please any suggestions??
Many thanks
Created 02-15-2016 12:09 PM
@Roberto Sancho please use Hive 1.2.1 and pig 0.15 jars.
This script uses Hive, pig and hcatalog and tested on latest sandbox https://github.com/dbist/oozie/tree/master/apps/hcatalog also execute pig with -use hcatalog switch
Created 02-15-2016 12:09 PM
@Roberto Sancho please use Hive 1.2.1 and pig 0.15 jars.
This script uses Hive, pig and hcatalog and tested on latest sandbox https://github.com/dbist/oozie/tree/master/apps/hcatalog also execute pig with -use hcatalog switch
Created 02-15-2016 12:13 PM
Please see this tutorial http://hortonworks.com/hadoop-tutorial/how-to-use-hcatalog-basic-pig-hive-commands/
You can follow it step by step to make this work.
Created 02-15-2016 12:13 PM
Many thanks, the pi version is :
Pig0.15.0.2.3
and the Hive version is:
Hive1.2.1.2
This is correct???
Thanks.
Created 02-15-2016 12:14 PM
Yes just look at my script @Roberto Sancho
Created 02-15-2016 12:28 PM
@Roberto Sancho as you see in my script I am not registering hive jars, just make sure you use -useHCatalog. I had same issue in this thread https://community.hortonworks.com/content/idea/2391/pig-and-hive-actions-should-include-hive-hcatalo...
also in your original script, you're mixing hive and pig 0.11 with hive 1.2.1. That's why I said use latest jars. If you still need to register, go to /usr/hdp/current/pig-client for pig jars and /usr/hdp/current/hive for hive jars. Remove 0.11 jars completely. But in my script I don't even register any pig or hive jars, just fix your classpath.
Created 02-15-2016 03:20 PM
Hi:
it work with me like this, many thanks 🙂
register /usr/lib/piggybank/hive-hcatalog-pig-adapter.jar register /usr/lib/piggybank/hive-common.jar; register /usr/lib/piggybank/hive-metastore.jar; register /usr/lib/piggybank/hive-exec.jar; register /usr/lib/piggybank/hive-serde.jar; register /usr/lib/piggybank/hive-shims.jar; register /usr/lib/piggybank/libfb303.jar