Member since
03-29-2020
110
Posts
10
Kudos Received
16
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
851 | 01-08-2022 07:17 PM | |
2689 | 09-22-2021 09:39 AM | |
12353 | 09-14-2021 04:21 AM | |
2337 | 09-01-2021 10:28 PM | |
2863 | 08-31-2021 08:04 PM |
09-21-2021
07:59 AM
Thanks @Shelton. Reading I found these limitations: - Replicating to and from HDP to Cloudera Manager 7.x is not supported by Replication Manager. * The only option I saw: - Use DistCp to replicate data. - Hive external tables. For information to replicate data, contact Cloudera Support =>? https://docs.cloudera.com/cdp/latest/data-migration/topics/cdpdc-compatibility-matrix-bdr.html
... View more
09-19-2021
06:03 AM
@Kiddo you either need to get the md5 of all the records or you could collect it for the concatenated form of all the column names in a table. e.g. select md5(concat(col1,col2)) from table1; or. select md5(concat(*)) from (SELECT collect_list(column_name) as col_name from information_schema.columns where table_schema='db1' and
table_name='table1')e; Hope this answers your question.
... View more
09-17-2021
06:56 AM
Hi @manojamr I am glad to know your original issue got resolved. As per your last comment, your Query took 9.5 hours to get complete. In this case, we may need to check whether there is a delay or hungriness, or resource crunch or it is normal. To figure out that we may need beeline console output, QueryId, Application log, all HS2 and HMS logs. It would be great if you create a case with Cloudera so we would be happy to assist you. If you are happy with the reply, mark it Accept as Solution
... View more
09-01-2021
11:44 PM
@Eric_B Yes, your understanding is correct.
... View more
09-01-2021
10:28 PM
1 Kudo
Hi @saikat As I can understand you are running a merge query and it is failing with java.lang.OutOfMemoryError error. Step 1: Could you please run major compaction on all the tables involves in the merge query(If it is an ACID table or else ignore step1). Once the major compaction is triggered make sure it got completed by running "show compactions;" command in the beeline. This will bring down some stats collection burden for the hive. How to run minor and major compaction? Alter table <table name> compact 'MAJOR'; Step 2: Once step1 is done. Please set the following propery in beeline session level and re-run the merge query set hive.tez.container.size=16384; set hive.tez.java.opts=-Xmx13107m; set tez.runtime.io.sort.mb=4096; set tez.task.resource.memory.mb=16384; set tez.am.resource.memory.mb=16384; set tez.am.launch.cmd-opts=-Xmx13107m; set hive.auto.convert.join=false; The TEZ container and AM size is set as 16GB, if the query got failed you can increase the value to 20GB(then hive.tez.java.opts and tez.am.launch.cmd-opts need to be configured 80% of container and AM size that is 16384). If the query got succeeded with 16GB of TEZ container and AM size then you can try to decrease it too 14/12/10 and figure out a benchmark where it is failing and getting succeeded. In this way, you can save resources. If you are happy with the comment, Mark it "Accept as Solution".
... View more
08-22-2021
12:09 AM
Hi @Nil_kharat Still not resolved the issue. You may need to check the HS2 logs and application logs to figure the slowness. And one more thing how can we track the job that are running by user's. 1. Go to RM UI > Running/finished/killed > check the User column 2. CM > YARN > Applications > Based upon the user you can search over here. If you are happy with the response mark it as Accepts as Solution
... View more
08-19-2021
12:59 AM
Hi @Nil_kharat Generally, in Hive you may see Query slowness, Query failure, Configuration issue, Alerts, Services down, Vulnerability issue, some bugs this kind of issues you may see.
... View more
08-14-2021
07:41 PM
Hi @Nil_kharat Yes, you can also use the particular lsof command as well to figure out the number of established connections to the particular port.
... View more
08-13-2021
04:50 AM
1 Kudo
Hi @amitshanker Thanks for the update. I can see than you had set hive.server2.webui.use.spnego as true, that means kerberos is enabled in your cluster. If spnego and kerberos are enabled then few settings needs to be changed in browser. Could you please follow the below link. https://docs.cloudera.com/documentation/enterprise/6/6.3/topics/cdh_sg_browser_access_kerberos_protected_url.html
... View more
08-12-2021
06:23 AM
Thanks it worked.
... View more