I am trying to custom parquet-avro schema creation for a table which is taken from Kafka Topic using Java Avro API. Output parquet file is worn with Hive table table. Decimal fields are created as fixed_len_byte_array on the schema. The Hive table can be queried using Hive and Spark. But, when impala query is used, it gives like below error.
I try to define byte array size according to cloudera impala documentation
1 <= precision <=9, then 4 bytes
10<= precision <=18, then 8 bytes
precision>18, then 16 bytes.
In the impala query screen on Cloudera Hue. It shows Hive table description.