Member since
07-20-2020
11
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
6968 | 07-23-2020 06:18 AM |
07-23-2020
06:18 AM
Since the solution is scattered across many posts, I'm posting a short summary of what I did. I am running HDP 2.6.5 image on VirtualBox. Increased my virtual hard disk through Virtual Media Manager In the guest OS, Partitioned the unused space Formatted the new partition as an ext4 file system Mounted the file system Update the /etc/fstab (I couldn't do it, as I did not find that file In Ambari, under DataNode directory config, added the newly mounted file system as a comma separated value Restarted HDFS (my cluster did not have any files, therefore I did not run the below) Thanks to @Shelton for his guidance. sudo -u hdfs hdfs balancer
... View more
07-22-2020
09:15 PM
increasing VDI size doesn't mean partitions/filesystems inside Guest VM will be increased - I figured this out. Anyway, I have been able to successfully increase the capacity of HDFS. I'm facing problems loading a 73GB file into hdfs. See my most recent post for more details (before this).
... View more
07-22-2020
08:01 PM
@aakulov @Shelton Using the hdfs dfs -put has given me the following error. Could it be because my virtual machine (and the hadoop cluster running on top of it), the VDI file all reside on a external hard drive? It is not an SSD and read/write speeds aren't extremely fast. I believe I can get better performance if everything ran off my machine's own internal hard drive (which is an SSD). Would using distcp help in this scenario?
... View more
07-22-2020
07:48 PM
Thank you for your inputs. I have been able to expand the size of my HDFS finally.
... View more
07-22-2020
09:06 AM
@Shelton So, I've been able to create a new partition and format it as an ext4 filesystem mount it How do I add add this new partition to my datanode? Is it as simple as putting the drive path in Amabari DataNode config?
... View more
07-21-2020
07:32 AM
Initially, I was unable to load data into my guest OS and therefore I used Amabari's GUI to load data into HDFS directly from my host OS. Now that I have figured out using WinSCP, I have been able to copy my file from the host OS to the guest OS. I have used the put command to copy files to my HDFS. However, I have run out of disk space on my virtual machine. I have increased my VDI size but still figuring out a way to ensure HDFS utilizes the newly allocated storage. Once I figure that out, I think loading the data won't be a problem anymore. I did read the documentation of distcp yesterday. Thank you for giving a summary here anyway. I may not be required to use distcp. Let me see how it goes.
... View more
07-21-2020
06:42 AM
It is a VDI. I have used Virtual Media Manager to increase the size of my disk. How can i get HDFS to expand and make use of the unallocated space? I'm assuming this is how one would do it 1. Create a new partition in the Guest OS and assign a mount point to it. 2. Add that path to the DataNode directories (or) Extend the current partition to fill the unused disk space so that DataNode automatically increases the HDFS size?
... View more
07-20-2020
08:15 PM
I don't think it is dynamically allocated, or at least it doesn't seem to be working. I've run out of space trying to load a ~70 GB file. How can I increase the capacity?
... View more
07-20-2020
04:04 PM
I'll look into it. I'll have to install gcc and then later Maven to run those shell scripts. Thanks for your input.
... View more
07-20-2020
04:02 PM
Hello, I'm trying to upload a 73 GB tsv file into HDFS through Ambari's File View option. The progress bar seems to get stuck at 3%-4% and doesn't proceed further. I've had to cancel the upload process after waiting for an hour to see the progress bar move. What can I do?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
-
HDFS