Member since
09-24-2015
144
Posts
72
Kudos Received
8
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1315 | 08-15-2017 08:15 AM | |
6156 | 01-24-2017 06:58 AM | |
1615 | 08-03-2016 06:45 AM | |
2914 | 06-01-2016 10:08 PM | |
2502 | 04-07-2016 10:30 AM |
03-02-2016
09:37 AM
1 Kudo
If you shutdown the OS all tasks running on that node will be stopped too so you don't need to worry about recovery. You might kill the running application masters on that node though. There is no graceful shutdown of a nodemanager that waits for running applications to finish as of yet ( AFAIK if someone knows better let me know ). Yarn depends on applications to handle task or AM failures gracefully. https://issues.apache.org/jira/browse/YARN-914
... View more
08-12-2016
05:31 AM
Now, to setup host and install HDP, only "./start_hdp.sh -a".
It automatically sets up the latest HDP in your Ubuntu 14.04 (16 is not supported)
To access, it starts a proxy on port 28080, so you can change the browser proxy setting to use Ubuntu_IP:28080.
Or, if Ubuntu and your PC are in same network, just adding a route to containers works (eg: route add -net 172.17.100.0/24 Ubuntu_IP on your Mac)
... View more
02-11-2016
01:23 AM
2 Kudos
In the capacity scheduler your could set a high priority queue at 90% of the cluster with extension( max. Capacity) to 100% and a low priority queue wirh 10% with extension( max capacity) to 100%. In this case jobs the first queue would always get 90% of the cluster if it needs it and the second queue would only get a tiny amount of the cluster if the high priority Queues have Queries. The low priority queue would still be able to monopolize the cluster if if has very long running tasks. But you could fix that with preemption. ( Or by making sure tasks in your cluster don't run for too long which they shouldn't anyway.)
... View more
02-02-2016
07:42 PM
1 Kudo
so... answer is it won't be fixed or middle of restoring process?
... View more
01-13-2017
12:59 AM
Just one question. AMBARI-12896 won't encrypt/obfscate password stored in ranger's xml file, will it?
... View more
12-16-2015
10:21 PM
Thank you! I will play with it.
... View more
12-10-2015
04:14 AM
1 Kudo
Hajime, the above scripts are for the yarn container and mapreduce memory settings. If you are trying to configure the memory of the nodemanager process itself then that shouldn't need more than 2GB - 4GB. If you are seeing outOfMemory there I suggest you turn on verbose GC for the nodemanager process and review the GC logs.
... View more
11-26-2015
10:33 AM
2 Kudos
@Hajime This makes sense hive.exec.reducers.bytes.per.reducer
Default Value: 1,000,000,000 prior to Hive 0.14.0; 256 MB ( 256,000,000 ) in Hive 0.14.0 and later Added In: Hive 0.2.0; default changed in 0.14.0 with HIVE-7158 (and HIVE-7917) Size per reducer. The default in Hive 0.14.0 and earlier is 1 GB, that is, if the input size is 10 GB then 10 reducers will be used. In Hive 0.14.0 and later the default is 256 MB, that is, if the input size is 1 GB then 4 reducers will be used. Point to note: Calculate hive.exec.reducers.max should be set to a number which is less than the available reduce slots on the cluster. Hive calculate the reducers based on hive.exec.reducers.bytes.per.reducer (default 1GB). Consider setting this high based on the workloads and demand for the reducers on the cluster https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties
... View more