Member since
01-25-2017
396
Posts
28
Kudos Received
11
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
830 | 10-19-2023 04:36 PM | |
4362 | 12-08-2018 06:56 PM | |
5456 | 10-05-2018 06:28 AM | |
19854 | 04-19-2018 02:27 AM | |
19876 | 04-18-2018 09:40 AM |
02-22-2017
03:22 PM
So i shouldn't search for the missing 40 T and the right storage is what the fsck shows?
... View more
02-22-2017
02:39 PM
CM 5.5.2 CDH 5.5.4
... View more
02-22-2017
07:03 AM
Hi, I changed the spark history server in my cluster before several weeks. I notice that when i'm clicking a finished spark job from cloudera manager, it's not directing me to the right spark history server url, while when i change the url manually i can see the finished the job.
... View more
Labels:
- Labels:
-
Apache Spark
02-21-2017
06:23 PM
Hi Lars, In your comments before you asked me to check the UID and i mentioned the version 'm using, unfortunately, it doesn't work.
... View more
02-20-2017
07:42 PM
Do you think looking at the edit logs size when this occur should be a good indication?
... View more
02-20-2017
07:09 PM
Hi, When i Run fsck on my cluster i got that several blocks under replicated and the target replication is 3 even i changed the dfs.replication to NN/ DNs and client server to replication factor 2, and mapred.submit.replication changed to 2. tried also: <property> <name>dfs.replication</name> <value>2</value> <final>true</final> </property> I also restarted all service at my cluster including the oozie. Looking at one of the running jobs conf and see the following with replication factor 3: mapreduce.client.submit.file.replication s3.replication kfs.replication dfs.namenode.replication.interval ftp.replication s3native.replication
... View more
02-20-2017
06:46 PM
The intersting thing that i noticed when this happened at the same time some jobs that runs once a day write to HDFS relatively too much data and it's run with a good number of reducers betweeb 400-1100, which make me suspect in the blocks that written by these jobs at the same time and the vm is getting some lag, trying to find a way to approve this.
... View more
02-20-2017
06:37 PM
Total size: 253714473531851 B (Total open files size: 11409372739 B) Total dirs: 1028908 Total files: 7639121 Total symlinks: 0 (Files currently being written: 107) Total blocks (validated): 8781147 (avg. block size 28893090 B) (Total open file blocks (not validated): 149) Minimally replicated blocks: 8781147 (100.0 %) Over-replicated blocks: 0 (0.0 %) Under-replicated blocks: 0 (0.0 %) Mis-replicated blocks: 0 (0.0 %) Default replication factor: 3 Average block replication: 2.8528664 Corrupt blocks: 0 Missing replicas: 0 (0.0 %) Number of data-nodes: 30 Number of racks: 1 FSCK ended at Mon Feb 20 21:33:23 EST 2017 in 190136 milliseconds The filesystem under path '/' is HEALTHY hadoop fs -du -s / 244412682417174 708603392967605 / hadoop fs -count -q / 9223372036854775807 9223372036846392726 none inf 987886 7395195 244417466380498 / the non HDFS reserved space is 10 GB and for 30 nodes so it's should not exceed 1 T with replication factor 3. it's really annoying.
... View more
02-19-2017
05:30 AM
hdfs dfs -du -h -s / 221.1 T 637.9 T / ==================== hdfs dfs -du -h -s . 204.2 M 1.2 G . ================= But in the UI i see it's 670 T I'm sure i'm missing something but cann't find it. Configured Capacity: DFS Used: Non DFS Used: DFS Remaining: DFS Used%: DFS Remaining%: Block Pool Used: 1.02 PB 670.54 TB 283.37 GB 368.96 TB 64.49% 35.48% 670.54 TB
... View more
Labels:
- Labels:
-
HDFS