Member since 
    
	
		
		
		09-21-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                85
            
            
                Posts
            
        
                75
            
            
                Kudos Received
            
        
                7
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 2487 | 04-21-2016 12:22 PM | |
| 5678 | 03-12-2016 02:19 PM | |
| 2431 | 10-29-2015 07:50 PM | |
| 2806 | 10-02-2015 04:21 PM | |
| 7621 | 09-29-2015 03:08 PM | 
			
    
	
		
		
		01-27-2021
	
		
		12:00 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi @abajwa,     Does the Ambari Server Host also need to present it's own SSL certificate to the AD server?  In case of multiple domain controllers, do we need to have separate SSL certificates from each of the domain controllers?     Thanks,  Megh 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		01-23-2020
	
		
		10:31 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Thanks it really fixes the problem RHEL 6.10 (Open JDK 1.8) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-09-2018
	
		
		12:45 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 The tutorial from which the script was linked has been deprecated on github, but there is a similar set of instructions here:https://docs.hortonworks.com/HDPDocuments/Cloudbreak/Cloudbreak-2.5.1/content/aws-launch/index.html 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-05-2018
	
		
		12:10 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 These seems to be bogus replay exception when running solr service.   Changes hadoop-env.sh or solr JVM option with -Dsun.security.krb5.rcache=none should fix the problem.  # # Extra Java runtime options.  Empty by default.
export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true -Dsun.security.krb5.rcache=none ${HADOOP_OPTS}"
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-23-2017
	
		
		11:10 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 The docs say "MariaDB 10" but RHEL7 comes with "MariaDB 5". 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		12-21-2016
	
		
		09:39 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Matt Foley - Does HDFS have home references such as ~ or ${HOME} ? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-10-2017
	
		
		12:34 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Thanks for the help.  I was able to run the view after setting the Hive session param to ssl=true;sslTrustStore=/var/log/ambari-server/hive-next-view/hive.jks;trustStorePassword=******** 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-21-2016
	
		
		01:37 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I concur with Sean. As long as any user, who have access to the cluster and the google personal key, they can explore GHFS bucket. I would say, google has to enhance the connector, by allowing intervention of kerberos prior to validation of the personal key. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		04-04-2016
	
		
		03:06 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		2 Kudos
		
	
				
		
	
		
					
							 Prerequisites: 
 
 Launch Sandbox on Azure
 
 VM Size: Minimum of A4 or A5 
 
 
 A Twitter App
 
 You'll use the API credentials 
 The "Application Details" don't matter 
 
 
 
 Prepare the Sandbox 
 Connect to SSH & Ambari 
 
 Connect to the Sandbox using SSH
 
 or web console: http://<<ip>>:4200/ 
 
 
 Become root:
 sudo su - 
 
 Reset the Ambari password:
 ambari-admin-password-reset 
 
 Login to Ambari:
 
 http://<<ip>>:8080 
 User: admin 
 
 
 Before moving to the next steps, ensure all services on the left are started (green) or in maintenance mode (black). 
 
 Install NiFi 
 
 In Ambari, Click "Actions" (bottom left) -> Add Service 
 Choose NiFi and continue through the dialogs. 
 You shouldn't need to change anything 
 NiFi should now be accessible at http:<<ip>>:9090/nifi/ 
 
 Tune Sandbox 
 The Sandbox is tuned to run on minimal hardware. We need to update the Hive, Tez & YARN configuration for our use case. 
 
 This could take up to 15 minutes to complete:
 bash <(curl -sSL https://git.io/vVRPs) 
 
 
 Solr & Banana 
 Solr enables the ability to search across large corpuses of information through specialized indexing techniques. 
 Banana is a dashboard visualization tool for Solr. 
 
 Download the Banana Dashboard
 curl -L https://git.io/vVRP3 -o /opt/hostname-hdpsearch/solr/server/solr-webapp/webapp/banana/app/dashboards/default.json 
 
 Update Solr to support Twitter's timestamp format
 curl -L https://git.io/vVRPz -o /opt/hostname-hdpsearch/solr/server/solr/configsets/data_driven_schema_configs/conf/solrconfig.xml 
 
 Start Solr
 JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk.x86_64 /opt/hostname-hdpsearch/solr/bin/solr start -c -z localhost:2181 
 
 Create Solr collection for tweets
 /opt/hostname-hdpsearch/solr/bin/solr create -c tweets -d data_driven_schema_configs -s 1 -rf 1
 
 
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		03-12-2016
	
		
		02:19 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		17 Kudos
		
	
				
		
	
		
					
							 As with many topics, "it depends".  For slave/worker/data hosts which only have distributed services you can likely disable swap. With distributed services it's preferred to let the process/host be killed rather than swap. The killing of that process or host shouldn't affect cluster availability.  Said another way: you want to "fail fast" not to "slowly degrade". Just 1 bad process/host can greatly degrade performance of the whole cluster. For example, in a 350 host cluster removal of 2 bad nodes improved throughput by ~2x:   http://www.slideshare.net/t3rmin4t0r/tez8-ui-walkthrough/23  http://pages.cs.wisc.edu/~thanhdo/pdf/talk-socc-limplock.pdf   For masters, swap is also often disabled though it's not a set rule from Hortonworks and I assume there will be some discussion/disagreement. Masters can be treated somewhat like you'd treat masters in other, non-Hadoop, environments.  The fear with disabling swap on masters is that an OOM (out of memory) event could affect cluster availability. But that will still happen even with swap configured, it just will take slightly longer. Good administrator/operator practices would be to monitor RAM availability, then fix any issues before running out of memory. Thus maintaining availability without affecting performance. No swap is needed then.  Scenarios where you might want swap:   playing/testing functionality, not performance, on hosts with very little RAM so will likely need to swap.  if you have the need to use more memory, or expect to need more, than the amount of RAM which has been purchased. And can accept severe degradation in failure. In this case you would need a lot of swap configured. Your better off buying the right amount of memory.   Extra thoughts:   if you want to disable swap, but your organization require their to be a swap partition, set swappiness=0  if you choose to have swap, set swappiness=1 to avoid swapping until all physical memory has been used.  most Cloud/Virtualization providers disable swap by default. Don't change that.  some advise to avoid swap on SSDs due to reducing their lifespan  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
         
					
				













