Member since 
    
	
		
		
		12-24-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                67
            
            
                Posts
            
        
                7
            
            
                Kudos Received
            
        
                0
            
            
                Solutions
            
        
			
    
	
		
		
		06-19-2017
	
		
		01:22 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 run the command ------> mount | grep /tmp      /dev/mapper/xxx_tmp on /var/tmp type xfs (rw,nosuid,nodev,noexec,relatime,attr2,inode64,noquota)      /dev/mapper/xxx_tmp on /tmp type xfs (rw,nosuid,nodev,relatime,attr2,inode64,noquota)     if you see noexec permission, remount with option.      sudo mount -o remount,exec /tmp/ 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-16-2017
	
		
		10:40 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Fixed after supplying temp path explicitly.   pyspark --driver-java-options -Djava.io.tmpdir=/home/<user-home>/temp  –v    OR     spark-shell --driver-java-options -Djava.io.tmpdir=/home/<user-home>/temp  –v 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		06-15-2017
	
		
		06:48 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 Getting below error as soon as submitting any command over spark-shell.        warning: there were 1 deprecation warning(s); re-run with -deprecation for details  [Stage 0:> (0 + 0) / 2]java.lang.reflect.InvocationTargetException  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)  at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)  at java.lang.reflect.Method.invoke(Method.java:606)  at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:312)  at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)  at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)  at org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)  at org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)  at org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)  at org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)  at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)  at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)  at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)  at java.lang.reflect.Constructor.newInstance(Constructor.java:526)  at org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)  at org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)  at org.apache.spark.broadcast.TorrentBroadcast.org$apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:74)  at org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:81)  at org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)  at org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)  at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1334)  at org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1006)  at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)  at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1611)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1603)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1592)  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)  Caused by: java.lang.UnsatisfiedLinkError: /tmp/snappy-1.0.4.1-libsnappyjava.so: /tmp/snappy-1.0.4.1-libsnappyjava.so: failed to map segment from shared object: Operation not permitted  at java.lang.ClassLoader$NativeLibrary.load(Native Method)  at java.lang.ClassLoader.loadLibrary1(ClassLoader.java:1965)  at java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1890)  at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1851)  at java.lang.Runtime.load0(Runtime.java:795)  at java.lang.System.load(System.java:1062)  at org.xerial.snappy.SnappyNativeLoader.load(SnappyNativeLoader.java:39)  ... 29 more  org.apache.spark.SparkException: Job aborted due to stage failure: Task serialization failed: java.lang.reflect.InvocationTargetException  sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)  sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)  sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)  java.lang.reflect.Constructor.newInstance(Constructor.java:526)  org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)  org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)  org.apache.spark.broadcast.TorrentBroadcast.org$apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:74)  org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:81)  org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)  org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)  org.apache.spark.SparkContext.broadcast(SparkContext.scala:1334)  org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1006)  org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)  org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)  org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1611)  org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1603)  org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1592)  org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)  at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1433)  at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1421)  at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1420)  at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)  at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1420)  at org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1016)  at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)  at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1611)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1603)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1592)  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1862)  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1875)  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1888)  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1959)  at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)  at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)  at org.apache.spark.rdd.RDD.collect(RDD.scala:926)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$.mergeSchemasInParallel(ParquetRelation.scala:799)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$MetadataCache.org$apache$spark$sql$execution$datasources$parquet$ParquetRelation$MetadataCache$$readSchema(ParquetRelation.scala:517)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$MetadataCache$$anonfun$12.apply(ParquetRelation.scala:421)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$MetadataCache$$anonfun$12.apply(ParquetRelation.scala:421)  at scala.Option.orElse(Option.scala:257)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$MetadataCache.refresh(ParquetRelation.scala:421)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation.org$apache$spark$sql$execution$datasources$parquet$ParquetRelation$$metadataCache$lzycompute(ParquetRelation.scala:145)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation.org$apache$spark$sql$execution$datasources$parquet$ParquetRelation$$metadataCache(ParquetRelation.scala:143)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$$anonfun$6.apply(ParquetRelation.scala:202)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation$$anonfun$6.apply(ParquetRelation.scala:202)  at scala.Option.getOrElse(Option.scala:120)  at org.apache.spark.sql.execution.datasources.parquet.ParquetRelation.dataSchema(ParquetRelation.scala:202)  at org.apache.spark.sql.sources.HadoopFsRelation.schema$lzycompute(interfaces.scala:636)  at org.apache.spark.sql.sources.HadoopFsRelation.schema(interfaces.scala:635)  at org.apache.spark.sql.execution.datasources.LogicalRelation.<init>(LogicalRelation.scala:39)  at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:125)  at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:109)  at org.apache.spark.sql.SQLContext.load(SQLContext.scala:1126)  at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:25)  at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:30)  at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:32)  at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:34)  at $iwC$$iwC$$iwC$$iwC.<init>(<console>:36)  at $iwC$$iwC$$iwC.<init>(<console>:38)  at $iwC$$iwC.<init>(<console>:40)  at $iwC.<init>(<console>:42)  at <init>(<console>:44)  at .<init>(<console>:48)  at .<clinit>(<console>)  at .<init>(<console>:7)  at .<clinit>(<console>)  at $print(<console>)  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)  at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)  at java.lang.reflect.Method.invoke(Method.java:606)  at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1045)  at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1326)  at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:821)  at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:852)  at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:800)  at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)  at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)  at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)  at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)  at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)  at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)  at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)  at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)  at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)  at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)  at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)  at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1064)  at org.apache.spark.repl.Main$.main(Main.scala:35)  at org.apache.spark.repl.Main.main(Main.scala)  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)  at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)  at java.lang.reflect.Method.invoke(Method.java:606)  at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:730)  at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)  at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)  at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)  at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)  Caused by: java.lang.reflect.InvocationTargetException  at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)  at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)  at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)  at java.lang.reflect.Constructor.newInstance(Constructor.java:526)  at org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)  at org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)  at org.apache.spark.broadcast.TorrentBroadcast.org$apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:74)  at org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:81)  at org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)  at org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)  at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1334)  at org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1006)  at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)  at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1611)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1603)  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1592)  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)  Caused by: java.lang.IllegalArgumentException: org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null  at org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:171)  at org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)  at org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)  at org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)  ... 18 more  Caused by: org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null  at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)  at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)  at org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)  ... 21 more    scala> 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Spark
			
    
	
		
		
		11-08-2016
	
		
		03:48 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 We have three journal nodes. One was creating issue of sync with name node. Even after multiple rolling restart did not fix sync issue. 
   
 Later we moved, Journal node services to another node but we again got sync issue with new host’s - journal node to name node. 
   
 Another things observed here is, location /Data0/dfs/jn directory was empty on new journal node host after moving one journal node service there. 
   
 Restart of Zookeeper & Journal node did not enable creation of edit logs on new host as well. We have also tried to move everything from working/old JN's location /Data0/dfs/jn/hanameservice/* to new host as per below section. 
 https://www.cloudera.com/documentation/enterprise/5-4-x/topics/cm_mc_jn.html 
   
 Can you please suggest, how to enable sync of Journal node with namenode or what step actually we are missing here? 
   
 CDH - 5.4.+ 
   
   
 Regards, 
 Kartik 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Apache Zookeeper
- 
						
							
		
			Cloudera Manager
			
    
	
		
		
		06-07-2016
	
		
		08:17 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							    Can cloudera director create external RDS which is encrypted using KMS ?     Can cloudera director create EBS volumn which is encrypted ?        Regards,  Kartik Bhatnagar 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-26-2016
	
		
		11:59 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 I have three questions on instance storage on ec2 instance.     1. How to give different volume for workers and master, if using director client ?      rootVolumeSizeGB: <??>     2. is the above root volumne added is of EBS kind ?     3. For example, If i select worker machine of type d2.2xlarge. what is use of  OOTB instance storage attached to EC2 volume since that is never mounted for root expansion ?    d2.2xlarge ===> 6 x 2048 GB (Use ?) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-24-2016
	
		
		08:46 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 My setup is done using private and public subnet using cloudformation scripts of cloudera director client.  Using below tunnel, i am able to use cloudera manager on localhost:7180  ssh -i my-key.pem -L 7180:<private IP of master>:7180 -L 7187:<private IP of master>:7187 ec2-user@<Public Ip of launcher(director client)>     I am able to use cloudera manager on localhost:7180.     how to access other services  like application master resource manager which are running on different port like 8088.     I am failed to access below kind of url, example.  http://ip-10-0-1-106.ec2.internal:8088/cluster/app/application_1463741301226_0001       
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
- 
						
							
		
			Cloudera Manager
			
    
	
		
		
		05-22-2016
	
		
		12:55 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Thanks Jayita for sharing HA Git hub link.     https://github.com/cloudera/director-scripts/blob/master/configs/aws.ha.reference.conf     Below is some part copied form above conf file.     masters-1 {    count: 2         instance: ${instances.i22x} {    tags {    group: masters-1    }    }         roles {    ZOOKEEPER: [SERVER]    HIVE: [HIVESERVER2, HIVEMETASTORE]    YARN: [RESOURCEMANAGER]    HBASE: [MASTER]    HUE: [HUE_SERVER]    OOZIE: [OOZIE_SERVER]    }         is it correct to have two HS2 and HMS running services at two masters since count is 2 so it will be started at two places.        Regards,  Kartik 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		05-18-2016
	
		
		10:03 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							    Hi Jayita,     Can you please also add below point.      HDFS availability can be accomplished by deploying the NameNode with high availability with at least three JournalNodes.      Regards,  Kartik Bhatnagar 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		 
        







