Member since
02-03-2018
5
Posts
0
Kudos Received
0
Solutions
01-21-2021
10:43 AM
I discovered my issue, I was missing the join statement for the queued inserts. Very basic issue, thank you for your time!
... View more
01-21-2021
10:10 AM
Hey Cloudera Community, We have a production DB that tends to write multiple small parquet tables to most tables. We wrote a consolidation script that creates a temp table per production table, copies the data over to the temp table, then insert overwrites back to the main table, then drop the temp table. Unfortunately one of our production tables was missed by this script and has far too many parquet files. The query details for trying to copy the whole table at once to the temp table indicate that Impala wants to allocate significantly more resources than are available. As a result, we decided to try to attempt iteratively selecting from our main table and inserting into the temp table (partitioned by a customerid and year). We have a variable number of threads, and a queue full of tuples containing the tenantids and years. Ideally each thread gets a tuple from the queue and then runs the insert statement, then repeats while the queue is not empty. When we try to run our script, we get a generic failure in Impala and the session is closed. Does anybody know I'm doing incorrectly?
... View more
Labels:
- Labels:
-
Apache Impala
08-28-2020
01:15 PM
I decreased the ram usage for our name node and bumped it up for the others and the problem went away.
... View more
08-28-2020
12:04 PM
Hello Cloudera Forums, IT recently cloned our VMs for one of our old Hadoop stacks onto some fancy new hardware and bumped up the specs on the VMs. We had been getting "out of memory" errors on Impala Daemon so IT provided us with a significant boost to the available RAM on each server (I know that we SHOULD have just added on another node but unfortunately that's just how my org operates). Anyways we modified the configuration for the services to use more of the RAM on the VMs but now we are getting the below error: Memory on host XXXX is overcommitted. The total memory allocation is 223.8 GiB bytes but there are only 252.3 GiB bytes of RAM (50.5 GiB bytes of which are reserved for the system). Visit the Resources tab on the Host page for allocation details. Reconfigure the roles on the host to lower the overall memory allocation. Note: Java maximum heap sizes are multiplied by 1.3 to approximate JVM overhead. I'm a bit confused as to why we are getting this error, when I increase the RAM used by Impala I still get this error (the system reserving 50Gb) but when I decrease it, the number doesn't change either. I'm a bit confused as to what the issue is. Is there a shared role across the nodes that needs a mirrored config or something?
... View more
Labels: