Member since 
    
	
		
		
		02-01-2019
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                650
            
            
                Posts
            
        
                143
            
            
                Kudos Received
            
        
                117
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 3502 | 04-01-2019 09:53 AM | |
| 1812 | 04-01-2019 09:34 AM | |
| 8868 | 01-28-2019 03:50 PM | |
| 1971 | 11-08-2018 09:26 AM | |
| 4461 | 11-08-2018 08:55 AM | 
			
    
	
		
		
		06-08-2017
	
		
		09:08 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @sai ram Here is the "Cancel Subscription"     
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-08-2018
	
		
		07:47 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 It has been a while and I believed you already got it working. But for those people want to know how to do it, I am going to show you how I did it.  Before I get started, I wanted to let you know I found the informative  link  by googling "setup Jupyter notebook at Hortonworks sandbox". Based on the link and I made some minor changes, I got it working.  ####========================================================
### login as root
####========================================================
sandbox-version
== Sandbox Information ==
Platform: hdp-security
Build date: 06-18-2018
Ambari version: 2.6.2.0-155
Hadoop version: Hadoop 2.7.3.2.6.5.0-292
OS: CentOS Linux release 7.5.1804 (Core)
====
####========================================================
### Install Jupyter Dependencies
####========================================================
pip install --ignore-installed pyparsing
yum install epel-release
sudo wget https://bootstrap.pypa.io/ez_setup.py -O - | python ;sudo yum install python-pip python-wheel python-devel gcc
pip install --upgrade pip
pip install --upgrade pip wheel pandas numpy scipy scikit-learn matplotlib virtualenv
####========================================================
### Install Jupyter
####========================================================
pip install jupyter
####========================================================
### Setup folders and files
####========================================================
jupyter notebook --generate-config
sudo mkdir -p /ibm/conf
sudo chown -R spark:hadoop /ibm
cp ~/.jupyter/jupyter_notebook_config.py /ibm/conf/
####========================================================
### Setup startup shell script
####========================================================
vi /ibm/scripts/start_jupyter.sh
#copy the paste the following contents
#! bin/bash
set -x
USER=$1
JUPYTER_HOST=sandbox-hdp.hortonworks.com
JUPYTER_PORT=8889
su - ${USER} << EOF 
export SPARK_HOME=/usr/hdp/current/spark-client
export PYSPARK_SUBMIT_ARGS="--master yarn-client pyspark-shell"
export HADOOP_HOME=/usr/hdp/current/hadoop-client
export HADOOP_CONF_DIR=/usr/hdp/current/hadoop-client/conf
export PYTHONPATH="/usr/hdp/current/spark-client/python:/usr/hdp/current/spark-client/python/lib/py4j-0.9-src.zip"
export PYTHONSTARTUP=/usr/hdp/current/spark-client/python/pyspark/shell.py
export PYSPARK_SUBMIT_ARGS="--master yarn-client pyspark-shell"
echo "Starting Jupyter daemon on HDP Cluster ..."
jupyter notebook --config=/ibm/conf/jupyter_notebook_config.py --ip=${JUPYTER_HOST} --port=${JUPYTER_PORT}&
EOF
exit 0
####========================================================
### Run startup shell script
####========================================================
chown -R spark:hadoop /ibm
chmod 777 /ibm/script/start_jupyter.sh
cd /ibm/scripts
./start_jupyter.sh spark
####========================================================
### Copy the link from above step's output and paste to your computer's browser 
####========================================================
# make sure you define sandbox.hortonworks.com in your hosts file
http://sandbox.hortonworks.com:8889/?token=c982c0f95222abcf2900e3aeb9d9c59cc0386cc04c6c154d   Test in Jupyter.     
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		07-19-2018
	
		
		03:23 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 hello, @Sandeep Nemuri  @Rahul Pathak   i have an issue with atlas i can see my data base neither my tables all my services are on   Any suggestion any ideas whitch configuration  should do   I really need help thanks you  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		12-05-2017
	
		
		12:24 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 try ssh -p2222 root@xx.xx.xx.xx' the  allocated IP    then it will ask where you want to add  say yes  change unix password  then run command 'ambari-admin-password-reset'  it will allow you to change the password for ambari admin 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-29-2017
	
		
		03:03 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 You have to login to HBase and remove master table atlas_titan as below  And restart service.  hbase(main):003:0> list  TABLE  ATLAS_ENTITY_AUDIT_EVENTS  atlas_titan  2 row(s) in 0.0070 seconds  => ["ATLAS_ENTITY_AUDIT_EVENTS", "atlas_titan"]  hbase(main):005:0> disable 'atlas_titan'  0 row(s) in 2.5060 seconds  hbase(main):006:0> drop 'atlas_titan'  0 row(s) in 1.2730 seconds  hbase(main):007:0> exit  Restart Atlas service from Ambari UI 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-31-2018
	
		
		11:53 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi all,  I have a similar issue, all 3 node managers in my cluster are marked as bad with local dirs bad alerts and I can assure I have plenty of space on the YARN disk. (just 35% of the disk is used). Does anyone know of this alert other than YARN reaching its disk threshold value??    Thanks,  Priyanshu  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-22-2016
	
		
		02:18 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							@bigdata.neophyte  -t option seems to not available in 2.7.1 version which is included with HDP 2.4.2.   This option is included in hadoop 2.7.3 which is available with HDP 2.5 Tech preview .   [hdfs@hdp1 ~]$ hadoop version
Hadoop 2.7.3.2.5.0.0-1133
Subversion git@github.com:hortonworks/hadoop.git -r 93bf28063ef319be6833d3d6f117d44e0b6b8fa9
Compiled by jenkins on 2016-08-03T11:38Z
Compiled with protoc 2.5.0
From source with checksum 1aed9e48ca6f7cd4ada3a36b1cd5feb
This command was run using /usr/hdp/2.5.0.0-1133/hadoop/hadoop-common-2.7.3.2.5.0.0-1133.jar
[hdfs@hdp1 ~]$ hdfs dfs -ls -t -help
-ls: Illegal option -help
Usage: hadoop fs [generic options] -ls [-C] [-d] [-h] [-q] [-R] [-t] [-S] [-r] [-u] [<path> ...]
[hdfs@hdp1 ~]$ hdp-select versions
2.5.0.0-1133
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		08-16-2016
	
		
		09:58 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 @sujitha sanku : "hadoop" is the root password of mysql server in HDP 2.5 sandbox. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		07-02-2016
	
		
		02:32 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 user name will be "admin" before this you need to reset the password form command line using : ambari-admin-password-reset   
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-25-2016
	
		
		01:21 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 @Ayub Pathan  These issues still exist. And I post the descriptions in the last answer, please check/
  
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		- « Previous
- Next »
 
        













