Member since
06-07-2016
923
Posts
322
Kudos Received
115
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3993 | 10-18-2017 10:19 PM | |
4255 | 10-18-2017 09:51 PM | |
14636 | 09-21-2017 01:35 PM | |
1773 | 08-04-2017 02:00 PM | |
2358 | 07-31-2017 03:02 PM |
02-27-2017
05:07 PM
1 Kudo
@Joe Petro What's the username for the user starting Nifi? What are the permissions for /var/lib/ambari-agent/cache/common-services/NIFI/1.0.0/package/files/nifi-toolkit-1.1.0.2.1.1.0-2/bin/encrypt-config.sh
... View more
02-27-2017
05:01 PM
@Akash S Are you seeing any performance issues with this? If not, then this is just your Garbage collection working the way it does. In fact it may not report how old generation objects are collection correctly. Your heap is also large (normal for hadoop -not for Java), so after a while it runs "stop the world" garbage collection and you see heap usage going down. This is likely not an issue but just to dig deeper, which Garbage collector are you using?
... View more
02-27-2017
04:35 PM
1 Kudo
@Sedat Kestepe There shouldn't be any issue. Network boning is done all the time to increase bandwidth as well as to provide redundancy.
... View more
02-24-2017
03:11 PM
@Aruna Sameera Did you install HBase on top of local file system? If not, then if its running on HDFS, then do you have /home/aruna/hbase-1.2.4/hbasestorage in your HDFS? Usually HBase runs on HDFS and the location for root directory is /hbase or /hbase/data
... View more
02-23-2017
08:06 PM
@Faruk Berksoz Here is the link to JDBC driver for phoenix. https://mvnrepository.com/artifact/org.apache.phoenix/phoenix-server-client/4.7.0-HBase-1.1
... View more
02-21-2017
02:50 PM
1 Kudo
@satya gaurav number of reducers is determined exactly by mapreduce.job.reduces. This is not just a recommendation. If you have specified a higher number of reducers, container allocation is still done based on queue size for that application. This is determined by your scheduler. Just because you request more than you should doesn't mean that those resources will be allocated. Your reducers will be waiting in queue until other complete. To get more details, you need to understand schedulers (capacity scheduler to be precise). Minimum container size is given by yarn.scheduler.minimum-allocation-mb (your request for less than this value will still result in a container with this minimum value and not a value you specify if its less than this). Similarly there is an upper limit given by yarn.scheduler.maximum-allocation-mb. Guess what happens if you request more than this? You don't get it. You get assigned this value if you request memory more than this. There are similar settings for core. This is at the cluster level. For each node, containers are allocated by Node Manager which of course is asking Resource Manager to do its job. yarn.nodemanager.resource.memory-mb is how much memory a container will allocate and yarn.nodemanager.resource.cpu-vcores is for CPU
... View more
02-21-2017
06:55 AM
@satya gaurav Check the following link. I don't want to repost the items as is, but the explanation here is what you are looking for: http://www.bigdatanews.com/profiles/blogs/hadoop-yarn-explanation-and-container-memory-allocations
... View more
02-21-2017
06:38 AM
1 Kudo
@satya gaurav Number of reducers is determined by mapreduce.job.reduces. Typically set to 99% of the cluster's reduce capacity, so that if a node fails the reduces can still be executed in a single wave. As for Application Master, you first need to understand YARN components. There is resource manager which has two main components: Scheduler and Application Manager (NOT Application Master). The ApplicationsManager is responsible for accepting job-submissions, negotiating the first container for executing the application specific ApplicationMaster and provides the service for restarting the ApplicationMaster container on failure. The per-application ApplicationMaster has the responsibility of negotiating appropriate resource containers from the Scheduler, tracking their status and monitoring for progress. Question: suppose if there is 2gb ram is available in a nodemanager and we submitted a job with 3gb ram then how job will run or it will not run. You need to understand how capacity scheduler works. There are assigned queues for the application(job) and capacity scheduler guarantees certain number of resources available to that application. If resources are available from other queues then your job can borrow those resources. This means if the queue is configured to use 2GB but if your job needs more resources, it can borrow. Please see the section titled "YARN Walkthrough" on following page: http://hortonworks.com/blog/apache-hadoop-yarn-concepts-and-applications/ screen-shot-2017-02-21-at-121739-am.png
... View more
02-20-2017
10:09 PM
@Raja Sekhar Chintalapati Do you have all Kerberos tickets generated and valid? This is an authentication error where you only need to generaqte right tickets if you are using Ambari.
... View more
02-19-2017
02:07 AM
@joseph c actually on a second thought, I realize 15-JAN-14 is also greater than 01-JUN-2011 because 15 is greater than 01 and it is less than 30-JUN-2011 because 15 is less than 30. HBase is behaving exactly how it is supposed to behave. You need to store date in a different format, like YYYYMMDD. When you do that, 20140115 does not fall between 20110601 and 20110630 lexicographically and you will get your expected result. Another way is to convert Date to timestamp.
... View more