Member since
07-31-2019
346
Posts
259
Kudos Received
62
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2938 | 08-22-2018 06:02 PM | |
1693 | 03-26-2018 11:48 AM | |
4217 | 03-15-2018 01:25 PM | |
5087 | 03-01-2018 08:13 PM | |
1434 | 02-20-2018 01:05 PM |
02-23-2016
09:08 PM
3 Kudos
@nasghar though you can export MS Access to a csv and import that into Hive I would suggest instead importing the data into SQL Server and use Sqoop.
... View more
02-22-2016
08:54 PM
1 Kudo
It's making more sense. My yarn.nodemanager.resource.memory-mb was only set to 16 GB so this restricted my min and max settings. Still not clear what to set disks to in a virtual environment in order to get a good baseline setting.
... View more
02-22-2016
08:28 PM
2 Kudos
We are running an 8 node virtualized cluster with 5 datanodes. Each datanode is allocated 8 vcores, 54 GB of RAM, and use shared SAN storage. The output of yarn-utils (v=8, m=54, d=4) is: yarn.scheduler.minimum-allocation-mb=6656 yarn.scheduler.maximum-allocation-mb=53248 yarn.nodemanager.resource.memory-mb=53248 mapreduce.map.memory.mb=6656 mapreduce.map.java.opts=-Xmx5324m mapreduce.reduce.memory.mb=6656
mapreduce.reduce.java.opts=-Xmx5324m yarn.app.mapreduce.am.resource.mb=6656 yarn.app.mapreduce.am.command-opts=-Xmx5324m mapreduce.task.io.sort.mb=2662 Some questions I have is 1) what do you put for disks value when data node disks are running on shared SAN storage? and; 2) The maximum container size only shows 8 GB even though each node is assigned 54 GB. Does this have something to do with over commitment in the virtual environment? yarn-utils wants it set to 53 GB.
... View more
Labels:
- Labels:
-
Apache YARN
02-18-2016
01:35 PM
2 Kudos
@Anshul Sisodia you may want to begin transitioning from Hue to Ambari Views. There is a File Browser view you can use to upload files.
... View more
02-17-2016
01:29 AM
2 Kudos
@Jeremy Salazar since the error states that the user is "ambari" you will need to add the following values to the HDFS custom core-site configuration: hadoop.proxyuser.ambari.groups=*
hadoop.proxyuser.ambari.hosts=* Once that's done you'll need to follow @Neeraj Sabharwal step and create your home directory and assign access.
... View more
02-09-2016
03:52 AM
2 Kudos
@teru mat make sure you create a SQL server user in SQL Server which matches the user specified in the HDP install. This will need to be a native SQL user. You can't use Windows authentication since HDP on Windows does not currently support Kerberos.
... View more
02-09-2016
03:44 AM
5 Kudos
@Sunile Manjee The Apache HDFS documentation states To minimize global bandwidth consumption and read latency, HDFS tries to satisfy a read request from a replica that is closest to the reader. I would guess that increasing the number of replicas increases the chances that the replica will reside close to the reader. Probably simplistic, but a logical guess.
... View more
02-08-2016
09:05 PM
1 Kudo
@Sunile Manjee though I have no personal experience with them there are companies like BlueData who abstract the storage component and provide a interesting private cloud experience based on containers. An interesting read on this subject is a book by Google called Datacenter as a Computer.
... View more
02-06-2016
03:07 PM
1 Kudo
@Malek Ben Salem was this resolved? If so, please either accept an answer or post the solution. Thanks!
... View more
02-05-2016
05:51 PM
1 Kudo
If you are using the sandbox on virtualbox, you'll need to add the Nifi port for port forwarding. 2016-02-05-11-52-14.png
... View more