Hi, I'm trying to create hive table and query it using %sql.
I tried both: %spark and %sql, nothing helps.
Here is my DDL:
%sql
create external table MY_TABLE row format serde 'com.my.MyAvroSerde'
with serdeproperties ('serialization.class'='com.my.ContainerProto')
stored as inputformat 'com.my.ProtoAvroFileFormat'
LOCATION 'hdfs://my/data'
Thrown exception:
MetaException(message:org.apache.hadoop.hive.serde2.SerDeException java.lang.ClassNotFoundException: Class com.my.ContainerProto not found)
It's confusing since spark paragraph works well with
code
%spark
import com.my.ContainerProto
// bla-bla
val rdd = sc.newAPIHadoopFile[AvroKey[ByteBuffer], NullWritable,
AvroKeyInputFormat[ByteBuffer]]("hdfs://my/data")
rdd.map{bytes => ContainerProto.fromBytes(bytes)}
Code executed and result produced. Why sql or spark paragraph doesn't see my 3rd party jars when I try to create hive table? Spark interpreter has proper configuration for required third party jars.