Member since
09-28-2018
8
Posts
0
Kudos Received
0
Solutions
07-31-2019
01:39 AM
Hello @bgooley , Thank you very much for your answers
... View more
07-24-2019
01:17 AM
Hi @bgooley, Thank you for your answer. Yes sorry, each time I mean Certificate Authority (CA). Ok then I will see what will be the best for us between our own CA or a public CA. And I will learn more about how to create and maintain a CA with your link. In fact our IT admins will be the only clients on CM and HUE. Our production cluster (private cloud) will generate outputs to another webportal service (public cloud) where there will be the final users. If I may ask, I have a last question to be sure: 1) Public CA won't work for host with private IP (private cloud) right ? Thank you in advance, Vincent
... View more
07-23-2019
09:00 AM
Hi all, We have currently a Hadoop cluster in test in a private cloud with private ip and hostname: - <private_ip_1> mh1.example.com (namenode 1 in private cloud) - <private_ip_2> mh2.example.com (namenode 2 in private cloud) - <private_ip_3> w1.example.com (worker 1 in private cloud) ... - <private_ip_n> wn.example.com (worker n private cloud) - <private_ip_m> eg1.example.com (edge node 1 + cloudera manager service in public cloud) - <private_ip_o> eg2.example.com (edge node 2 + hue etc. in public cloud) We can access on the private cloud only through the edge nodes and the cluster is kerberized with sentry. For the moment we use only self-certificate, but now we want to put the cluster in production. My question is that: Could we use official authority certificate in our case ? Because if I undestand well, authority certificate is only for public domain and public IP. Or what I was thinking for production is: - Self-Certificate for Cloudera Manager Agents (in all nodes in private cloud) - Authority Certificate for HUE UI on edge node on public cloud (with public ip adress) - Authority Certificate for Cloudera Manager UI on edge node on public cloud (with public ip adress) Do you think it's ok or overkill for production ? Am I missing something ? Thank you in advance, Vincent
... View more
Labels:
- Labels:
-
Cloudera Manager
10-22-2018
02:42 AM
Hi, No problem for the delay. Yes to resume, we have between 10 and 1000 messages per seconds to ingest indu Kudu, and each message is about 200+ bytes. And using Impyla we do individual row insertion (or insertion for 5 or 10 messages), does that explain all the small data blocks? Using CTAS it's much better thanks. But in general, do you have any recommandation for fast individual row insertion without too increasing memory usage? And in case of a slow streaming write ? The thing is that we would like to query the table fast enough with the latest data. Many thanks, Vincent
... View more
10-08-2018
05:53 AM
Hello, Thank you for your answer. Below is a scan RPC trace: "method_name": "kudu.tserver.ScanRequestPB", "samples": [ { "header": { "call_id": 1421646, "remote_method": { "service_name": "kudu.tserver.TabletServerService", "method_name": "Scan" }, "timeout_millis": 179999 }, "trace": "1008 12:09:55.490616 (+ 0us) service_pool.cc:163] Inserting onto call queue\n1008 12:09:55.490648 (+ 32us) service_pool.cc:222] Handling call\n1008 12:09:55.490657 (+ 9us) tablet_service.cc:1901] Found scanner 4f0b899b9c7d469a942854dfe3e1d921\n1008 12:09:55.494442 (+ 3785us) inbound_call.cc:157] Queueing success response\n", "duration_ms": 3, "metrics": [ { "key": "cfile_cache_miss", "value": 455 }, { "key": "cfile_cache_miss_bytes", "value": 402982 }, { "key": "lbm_reads_lt_1ms", "value": 455 }, { "key": "lbm_read_time_us", "value": 889 }, { "key": "spinlock_wait_cycles", "value": 25600 } ] }, Our worklow is relatively light for the moment: Kafka -> python consumer-> Kudu And the python consumer read about 25 messages per second with an average of 150 kb per second, and insert the data into kudu via impyla. (There are no updates and no deletes) Below is the output for the following commands on the tablet server (the 3 tablet servers have the same memory problem): sudo -u kudu kudu fs check --fs_wal_dir=<wal dir> --fs_data_dirs=<data dirs> We didn't expect to have so many missing blocks ... For a tablet with the command: sudo -u kudu kudu local_replica data_size 6a1b491538e24709808172aabd4cedae --fs_wal_dir=<wal dir> --fs_data_dirs=<data dirs>: And very often we couldn't use the command because of missing block: Finally we have also notice the "Write buffer memory usage" on tablets. It seems a bit high for our workload isn't it ? Thank you in advance, Vincent
... View more
10-05-2018
10:02 AM
Hello, Thank you for your answer. We have check, the block_cache_capacity_mb is 512 MiB. And below is an example of scan trace metrics : "trace": "1004 15:36:51.672528 (+ 0us) service_pool.cc:163] Inserting onto call queue\n1004 15:36:51.672553 (+ 25us) service_pool.cc:222] Handling call\n1004 15:36:51.673895 (+ 1342us) inbound_call.cc:157] Queueing success response\nRelated trace 'txn':\n1004 15:36:51.672634 (+ 0us) write_transaction.cc:101] PREPARE: Starting\n1004 15:36:51.672651 (+ 17us) write_transaction.cc:268] Acquiring schema lock in shared mode\n1004 15:36:51.672652 (+ 1us) write_transaction.cc:271] Acquired schema lock\n1004 15:36:51.672652 (+ 0us) tablet.cc:400] PREPARE: Decoding operations\n1004 15:36:51.672662 (+ 10us) tablet.cc:422] PREPARE: Acquiring locks for 1 operations\n1004 15:36:51.672666 (+ 4us) tablet.cc:426] PREPARE: locks acquired\n1004 15:36:51.672666 (+ 0us) write_transaction.cc:126] PREPARE: finished.\n1004 15:36:51.672674 (+ 8us) write_transaction.cc:136] Start()\n1004 15:36:51.672675 (+ 1us) write_transaction.cc:141] Timestamp: P: 1538667411672672 usec, L: 0\n1004 15:36:51.672694 (+ 19us) log.cc:582] Serialized 1538 byte log entry\n1004 15:36:51.673861 (+ 1167us) write_transaction.cc:149] APPLY: Starting\n1004 15:36:51.673876 (+ 15us) tablet_metrics.cc:365] ProbeStats: bloom_lookups=0,key_file_lookups=0,delta_file_lookups=0,mrs_lookups=0\n1004 15:36:51.673879 (+ 3us) log.cc:582] Serialized 25 byte log entry\n1004 15:36:51.673886 (+ 7us) write_transaction.cc:309] Releasing row and schema locks\n1004 15:36:51.673888 (+ 2us) write_transaction.cc:277] Released schema lock\n1004 15:36:51.673889 (+ 1us) write_transaction.cc:196] FINISH: updating metrics\n1004 15:36:51.673918 (+ 29us) write_transaction.cc:309] Releasing row and schema locks\n1004 15:36:51.673919 (+ 1us) write_transaction.cc:277] Released schema lock\n", But nothing seems to be anormal ? For the scanner we use impala via HUE, and when we check on Cloudera Manager the query state is "FINISHED" so I think that the scanner must be closed. Is there a way to track the block cache or how to refresh it? In fact we have removed this table and we are trying to reproduce the same situation. But we have noticed that when we fill the table, the tablet server memory keep increasing (slowly) over time. To fill the table, we have a loop open on Kafka and with impyla we insert new messages into Kudu. And once again, to get Kudu tablet server "default" memory we have to restart Kudu. We are trying to see if it's because we don't close impyla cursor, but are we missing something else? Thank you in advance, Vincent
... View more
10-01-2018
01:55 AM
Hi, Many thanks for your answer. Below is the schema of the table. The memory issue happens for any kind of queries, for example: -select * from table ORDER BY column_string ASC -select * from table WHERE column_string = 'abc' But even at rest, the RAM consumption is about 2.6/4G for this table of 700k rows (with a total tablet size of 8 G) it seems abnormal. Below are the links to the svg files: https://sendeyo.com/up/d/96b96e542a (Kudu Tablet Server at rest) https://sendeyo.com/up/d/752e5b1b98 (Kudu Tablet Server after a query) Thank you in advance, Vincent
... View more
09-28-2018
02:59 AM
Hi, We have done some successful tests on Kudu for a few months with the following configuration: Cluster Test - A *3 Kudu Masters *3 Tablet Servers *Sentry+Kerberos enabled in the cluster *Ubuntu 14.04 *Kudu 1.5.0 After that, we would like to put in production in our system Cluster Production - B *Same configuration in cluster A *Ubuntu 16.04 *Kudu 1.7 But we are currently experiencing memory errors we never had before on the Cluster A. After querying on a small table (700k rows and 30 columns), all the tablet servers have their memory full but the peak never goes back down, and we can't figure out where does it comes from. So we can't insert new rows etc. The only way we find to empty memory was to restart kudu ... Our memory configuration: memory_limit_hard_bytes = 4GB block_cache_capacity_mb = 512MB For example after a query, the memory of a tablet server: And in the log of a tablet server: We have also tried to debug memory usage by using gperftools, but a bit complicated to understand (I can send the svg files if needed) There are also a similar problem in the link below, but we didn't find any solutions https://issues.apache.org/jira/browse/KUDU-1762 Do you have any ideas ? Are we missing something ? Thank you in advance, Vincent
... View more
Labels:
- Labels:
-
Apache Kudu