Member since
03-03-2016
69
Posts
46
Kudos Received
0
Solutions
05-16-2018
12:59 PM
I had similar issue and fixed with the following ambari repo. wget -nv http://public-repo-1.hortonworks.com/ambari/centos6/2.x/updates/2.6.1.5/ambari.repo -O /etc/yum.repos.d/ambari.repo
... View more
12-26-2016
07:57 PM
1 Kudo
In your code, you may have to add the following line: conf.set("dfs.nameservices", "HadoopTestHA")
... View more
10-21-2016
05:26 PM
More details can be viewed from the "source" at http://hadoop.apache.org/docs/r2.7.3/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html
... View more
10-11-2016
04:19 PM
you might need to enable hadoop debug mode to get more visibility over the issue- export hadoop.root.logger=DEBUG and run the job from cli and test
... View more
09-29-2016
03:55 PM
1 Kudo
Windows-compatible HDP 2.4.3 is the last version of HDP that runs on Windows. It does not have a GUI to manage the cluster. Install instructions are here: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_installing_manually_book/content/meet-min-system-requirements.html HDP 2.3.x is deprecated (and for Windows did not have a GUI to manage the cluster). Linux only HDP 2.5+ does not install on Windows. Microsoft Cloud You can install HDP to Microsoft Azure cloud platform and this will install Ambari ... but the images are Linux. http://hortonworks.com/blog/easy-steps-to-create-hadoop-cluster-on-microsoft-azure/
... View more
10-05-2016
05:17 AM
No I am not looking for learning purpose but I am doing research for my organization to take this in production as well later. Thanks @Arun
... View more
09-23-2016
04:50 PM
6 Kudos
@Viraj Vekaria Yes. HDFS does the automated sharding and you have no control on it, but rarely one thinks about sharding of a file system like HDFS, but of an actual database like RDBMS, MongoDB or HBase. Semantics, but you asked to use the sharding in HDFS which implies manual control. It is done automatically. At most, what you could do is to change the global replication factor, change the replication factor per file, but you can't do anything about what is replicated where, no data locality control. Since @Justin Watkins mentioned traditional RDBMS and I mentioned also MongoDB and you asked about HDFS, I will summarize the differences in approach to achieve scalability between these three with the added HBase touch. Traditional RDBMS often run into bottlenecks with scalability and data replication when handling large amounts of data/data sets. There are some creative ways to setup master-slave setups to achieve some scalability and performance and all are coming by design, not out of box sharding. MongoDB sharding can be applied to allow distribution across multiple systems for horizontal scalability as needed. Like MongoDB, Hadoop’s HBase database accomplishes horizontal scalability through database sharding. Distribution of data storage is handled by the HDFS, with an optional data structure implemented with HBase, which allocates data into columns (versus the two-dimensional allocation of an RDBMS in columns and rows). Data can then be indexed (through use of software like Solr), queried with Hive, or have various analytics or batch jobs run on it with choices available from the Hadoop ecosystem or your choice of business intelligence platform. If any of the responses is helpful, please don't forget to vote and accept the best answer. Thanks.
... View more
03-24-2016
08:00 AM
1 Kudo
@Eyad Garelnabi, Yes this is for initial loading only as we have already windows file server with my DMS. So need to do that initial loading and than v'll transfer the flow of our app to HDFS tech. Bdw thnx for the suggestion, I'll look forward to it. 🙂
... View more
03-14-2016
09:40 AM
3 Kudos
@Viraj Vekaria See this thread https://community.hortonworks.com/questions/402/how-to-setup-high-availability-for-ambari-server.html
... View more