Member since
01-04-2018
4
Posts
1
Kudos Received
0
Solutions
07-05-2018
07:37 AM
One other thing. It looks like there were some issues with the Ubuntu OS and after switching over to Centos 7.5 the CDH 5.15 install ran without much issues. I have a question though, in the the install screens it has a Data Node configuration value: DataNode Data Directory dfs.data.dir, dfs.datanode.data.dir Comma-delimited list of directories on the local file system where the DataNode stores HDFS block data. Typical values are /data/N/dfs/dn for N = 1, 2, 3.... These directories should be mounted using the noatime option, and the disks should be configured using JBOD. RAID is not recommended. In JBOD mode say the server has 20 hard disks so each of the 20 disk will have 20 file mount points. I think we need to set this value to comma-delimited /data/1/dfs/dn, /data/2/dfs/dn, /data/3/dfs/dn....../data/20/dfs/dn . Now what happens if some of the data nodes have different number of JBOD disks say 20 disks in some and 10 disks in others. Since this is a global variable dfs.data.dir how does it allocate the 20 data directories in those data nodes with only 10 JBOD hard disks? Since there is no hostname defined in this variable to indicate different nunber of disks in different hosts. Also in future if new datanodes are added with different number of disks how is this specified while adding new data nodes? Thanks!
... View more
04-09-2018
01:58 AM
Hi, I don't think there is a solution to export hive / impala metadata directly into excel file. You can export metastore to sql dump : https://discuss.pivotal.io/hc/en-us/articles/115000104847-How-to-migrate-Hive-from-one-Hadoop-cluster-to-another- Then, convert the sql dump to csv file : https://blog.twineworks.com/converting-a-mysql-dump-to-csv-files-b5e92d7cc5dd
... View more