I follow the instruction on the link below to load csv data into a Phoenix data using MapReduce.
https://phoenix.apache.org/bulk_dataload.html
The loading was successful with a csv files of 4k rows. When I tried a file with millions of records, I kept getting errors of "illegal data" or "The data exceeds the max capacity for the data type" or " CSV record does not have enough values (has 13, but needs 14." At the end of the loading, I got a summary like
Job Counters
Failed map tasks=4
Killed reduce tasks=1
Launched map tasks=8
Launched reduce tasks=1
Not a single record is loaded. I know most of rows in the file are with the correct layout. Is there a way to ignore the layout error, or export rows of invalid layout?