04-18-2018 02:31 PM
In Impala 2.11 we actually capped the max batch_size setting. Before that you could set it to an arbitrarily high value, which could have strange consequences. It's still a bit of a use-at-your-own-risk setting since it can have consequences for memory consumption and performance.
The real fix for this would be https://issues.apache.org/jira/browse/IMPALA-1618. Setting batch_size is just a workaround that may or may not work for you.
09-27-2018 11:29 PM
Can you tell me the way to set the BATCH_SIZE for impala jdbc connection? I tried but it is not working for me.
09-29-2018 12:25 PM
When connecting to a server that is running Impala 2.0 or later, you can use the driver to apply configuration properties to the server by setting the properties in the connection URL.
10-01-2018 02:30 AM
I tried setting the Batch size in the connection URL but I didn't get any performance boost in the query fetching time. I have posted my usecase in the cloudera forum. Kindly answer my questions :
10-01-2018 05:03 AM
Please try to change all these 3 params:
TSaslTransportBufSize=4000; RowsFetchedPerBlock=60536; SSP_BATCH_SIZE=60536;
10-01-2018 05:19 AM - edited 10-01-2018 05:23 AM
As you suggested me to set
in the connection URL. I did the changes but i am getting these errors
java.sql.SQLException: [Simba][ImpalaJDBCDriver](500051) ERROR processing query/ statement. Error Code: 0, SQL state: TStatus(statusCode:ERROR_STATUS, sqlState:H Y000, errorMessage:Invalid query option: SSP_BATCH_SIZE ), Query: SET SSP_BATCH_SIZE=60536. at com.cloudera.hivecommon.api.HS2Client.executeStatementInternal(Unknow n Source) ~[Impala-JDBC-41-1.0.0.jar!/:na]
java.sql.SQLException: [Simba][ImpalaJDBCDriver](500051) ERROR processing query/statement. Error Code: 0, SQL state: TStatus(statusCode:ERROR_STATUS, sqlState:HY000, errorMessage:Invalid query option: TSaslTransportBufSize ), Query: SET TSaslTransportBufSize=4000.
Help me set up the property.
02-09-2019 04:47 AM
While IMPALA-1618 steel open and unresolved, I confirmed that this "workaround" is safe and efficient (I'm using it on a large scope and during more than 9 months) so that this is the only solution I find to solve or -get around- this big problem.
Hope that the main problem will be fixed ASAP.
Thanks for the remark.