Member since
02-08-2016
793
Posts
669
Kudos Received
85
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3064 | 06-30-2017 05:30 PM | |
3980 | 06-30-2017 02:57 PM | |
3302 | 05-30-2017 07:00 AM | |
3878 | 01-20-2017 10:18 AM | |
8396 | 01-11-2017 02:11 PM |
06-13-2016
05:03 AM
Many thanks! That works like a charm. Probably we can amend the documentation to include that it needs to be added to AMBARI_JVM_ARGS. BR, Rahul
... View more
06-10-2016
07:25 AM
I have 3 Journal Noeds in my cluster, but they don't seem to fail.
... View more
06-07-2016
04:37 PM
Thanks for the info.
... View more
06-02-2016
05:07 PM
It is a fresh installation with one additional Ranger Admin installed and configured with nginx to redirect to the working one.
... View more
01-15-2019
07:20 AM
After changing the hostname in the ambari configuration file restart the ambari server.
... View more
12-05-2016
01:47 AM
Upload the hdfs-replication-workflow.xml into hdfs will resolve this problem. i.e. create a hdfs directory /apps/data-mirroring/workflows and upload it
... View more
05-31-2016
06:05 AM
3 Kudos
@Sagar Shimpi this is nice post explaining how this property used by RMContainerAllocator while allocating a container upon request. http://johnjianfang.blogspot.in/2014/08/delay-scheduling-in-capacity-scheduling.html
... View more
05-30-2016
05:15 AM
@Mon key The best way is to find which are corrupted blocks using below command - hdfs fsck /path/to/corrupt/file -locations -blocks -files And then try to manually remove this using "hadoop rm -r </path>" to avoid dataloss. But still fsck does not remove good copied of data blocks.
... View more
05-27-2016
01:20 PM
1 Kudo
Problem Statement: When you try to execute GET call using ambari api to list/GET services, it usually gives error as shown below - # curl -u admin:admin -H "X-Requested-By: ambari" -X GET http://<AMBARI_SERVER_HOST>:8080/api/v1/clusters/<cluster_name>/services/
curl: (1) Protocol http not supported or disabled in libcurl
OR # curl -u admin:admin -H "X-Requested-By: ambari" -X GET “http://node1.example.com:8080/api/v1/clusters/HDP_TEST/services/“
curl: (1) Protocol “http not supported or disabled in libcurl
Solution: It’s a tipical curl_php error, but, the error response is not very, ehmmmm easy to deduce. It’s simple, surely there is an extra space before ‘http’, so check the CURLOPT_URL declaration, and search for this space, and then, delete it!!!!
Make sure the syntax of double quotes is correct. This also leads to the error.
... View more
Labels:
05-28-2016
12:22 AM
3 Kudos
Answers by @Sagar Shimpi and @Lester Martin look pretty good to me. Some further explanations:
How does snapshots help for Disaster Recovery? What are the best practices around using snapshots for DR purposes? Especially trying to understand when data is directly stored on HDFS, Hive data and HBase data If you're using the current distcp for DR (i.e., using distcp copying data from one cluster to your backup cluster), you have an option to utilize snapshot to do incremental backup so as to improve the distcp performance/efficiency. More specifically, you can choose to take snapshots in both the source and the backup cluster and use -diff option for the distcp command. Then instead of blindly copying all the data, the distcp will first compute the difference between the given snapshots, and only copy the difference to the backup cluster. As I understand, no data is copied for snapshots, but only metadata is maintained for the blocks added/ modified / deleted. If that’s the case, just wondering what happens when the comamnd hdfs dfs -rm /data/snapshot-dir/file1 is run. Will the file be moved to the trash? If so, will the snapshot maintain the reference to the entry in trash? Will trach eviction has any impact in this case? Yes, if you have not skipped the trash, the file will be moved to the trash, and in the meanwhile, you can still access the file using the corresponding snapshot path. How does snapshots work along with HDFS quotas. For example, assume a directory with a quota of 1GB with snapshotting enabled. Assume the directory is closer to its full quota and a user deleted a large file to store some other dataset. Will the new data be allowed to be saved to the directory or will the operation be stopped because the quota limits have been exceeded? No, if the file belongs to the snapshot (i.e., the file was created before a snapshot was taken), you will not release quota by deleting it. You may have to delete some old snapshots or increase your quota limit. Also in some old hadoop versions you may find the snapshots also affect the namespace quota usage in a strange way, i.e., sometimes deleting a file can increase the quota usage. This has been fixed by the latest version of HDP.
... View more