Member since
11-19-2015
158
Posts
25
Kudos Received
21
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
11721 | 09-01-2018 01:27 AM | |
1096 | 09-01-2018 01:18 AM | |
3663 | 08-20-2018 09:39 PM | |
484 | 07-20-2018 04:51 PM | |
1461 | 07-16-2018 09:41 PM |
02-21-2022
04:33 PM
What command did you use to run this and what is the location of connect-distributed.properties location in hortonworks? Thanks, Anand
... View more
04-03-2020
10:54 AM
I have taken the (now broken) EsharEditor Repo and made it work for HDP2 & 3, with Hue 3 & 4, and bundled into an ambari management pack. Details here: https://github.com/steven-dfheinz/dfhz_hue_mpack
... View more
11-15-2019
03:49 AM
Hello @JordanMoore @shashankvc @sihi_yassine , As per your requirement you want list of all external hive tables along with HDFS path location. Database name,Table name,Table Type(External) and HDFS Location of hive External tables. First login to Hive metastore and use the database which is storing all hive metadata Use 3 tables TBLS,DBS and SDS tables , on top of this 3 tables we can apply joins on DB_ID and SD_ID For more information and query output please check below link . https://askdoubts.com/question/how-to-find-out-list-of-all-hive-external-tables-and-hdfs-paths-from-hive-metastore/#comment-19 Thanks, Mahesh
... View more
11-13-2019
03:21 AM
@JordanMoore I am getting this error java.lang.ClassNotFoundException: io.confluent.connect.storage.StorageSinkConnectorConfig when trying to add connector using REST api. I am following this documentation.
... View more
10-10-2018
12:40 PM
The Client can interact with the Hive in the below three ways:-
ü Hive Thrift Client: The Hive server is exposed as thrift service. Hence it is possible to interact with HIVE with any programming language that supports thrift.
ü JDBC Driver: Hive uses pure Type 4 JDBC driver to connect to the server which is defined in org.apache.hadoop.HIVE.JDBC.HiveDriver class. . Pure Java applications may use this driver in order to connect to application using separate host and port.
The BeeLine CLI uses JDBC Driver to connect to the HIVE Server.
ü ODBC Driver: An ODBC Driver allows application that support ODBC to connect to the HIVE server. By default Apache does not ships the ODBC Driver but it is freely available by many vendors.
... View more
10-06-2018
06:41 PM
It is a Java class for reading Hadoop SequenceFIles http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/input/SequenceFileInputFormat.html
... View more
10-03-2018
10:36 PM
Hello @nicole wells Please find the question that you copied (and my answer) here https://stackoverflow.com/a/51687883/2308683
... View more
09-20-2018
10:54 AM
Btw, they spam us with granular & top-notch resources. I think it worths the spam. ^.^
... View more
09-19-2018
06:22 AM
Hi Joradn , sorry for the delayed response. You can create a ZK host configuration group and add 3 more servers to it , but in this case all the servers will be running in the single ensemble which is managed by Ambari and any point of time there will be only leader and rest of servers will be follower only. You can check this by running bellow command on all the ZK server host "echo stat | nc localhost 2181 | grep Mode" as the command." Cheers Sumit
... View more
09-04-2018
06:58 PM
@Manish
Tiwari, perhaps you can look at https://docs.hortonworks.com/HDPDocuments/Cloudbreak/Cloudbreak-2.7.1/content/data-lake/index.html Otherwise, you can search https://docs.hortonworks.com/ for the keywords you are looking for
... View more
09-30-2018
01:31 AM
You only need to use a Schema Registry if you plain on using Confluent's AvroConverter Note: NiFI can also be used to do CDC from MySQL https://community.hortonworks.com/articles/113941/change-data-capture-cdc-with-apache-nifi-version-1-1.html
... View more
08-20-2018
09:43 PM
Thank you very much for your explanation and guidance, Jay! Unfortunately, sandbox.service is not listed in HDF (just sandbox-ambari-services.service and sandbox-splash-web-server.service) nor does the /root/start_scripts/start_sandbox.sh file exist as mentioned in the article. Any suggestions on where to find documentation for HDF 3.1 (Linux sandbox-hdf.hortonworks.com 4.17.2-1.el7.elrepo.x86_64 #1 SMP Sat Jun 16 11:18:11 EDT 2018 x86_64 x86_64 x86_64 GNU/Linux)?
... View more
07-24-2018
06:26 PM
Contrary to answer by @Harshali Patel, exhaustion is not defined as an uneven distribution, it is rather a cause of it. A datanode has a property that you can set which defines a threshold of data must be reserved for the OS on that server. Once that limit is exceeded, the datanode process will stop and log an error telling you to delete some files from it. HDFS will continue to function with the other datanodes. The balancer can be ran to keep storage space healthy and even.
... View more
07-20-2018
04:52 PM
TaskTracker & JobTracker doesn't exist with YARN. The default replication factor is 3.
... View more
06-20-2018
06:55 PM
Please see previous question - https://community.hortonworks.com/questions/167618/how-to-specify-more-than-one-path-for-the-storage.html
... View more
05-21-2018
05:26 AM
Thank you so much for the help
... View more
04-09-2018
03:12 PM
Hello, it seems you duplicated this post. https://community.hortonworks.com/questions/183845/hdpcd-exam-issue.html Please remove this one.
... View more
03-30-2018
03:06 AM
@Geoffrey Shelton Okot - That requires installing sh module into the Ambari python installation, which I do not want to maintain. If I could do that, then I would use a Python Hive driver.
... View more
01-27-2018
10:11 AM
A small correction. It's introduced in Ranger 0.7 and policies should look like this: //HDFS
resource: path=/home/{USER}
user: {USER}
//Hive
resource: database=db_{USER}; table=*; column=*
user: {USER} where {USER} would substitute the user id of the currently logged in user.
... View more
01-17-2018
08:47 PM
Duplicate of https://community.hortonworks.com/questions/158942/ms-access-odbc-connection-to-hive.html ?
... View more
01-16-2018
06:46 AM
@Jordan Moore Not really relevant to the question but no this is not the point. The use case here is data export, where some clients have their own BI tools, processes and so on. They just need the data, csv in a zip file. Other clients do not have this in place and have a different access to this data.
... View more
01-02-2018
10:26 PM
Ambari itself doesn't know those disks are mounted until you edit the host configurations for HDFS/YARN and update the data directory configurations. The Ambari Alert check will run periodically to see if those configured disks are mounted, then the agent will update the dashboard.
... View more
12-01-2017
04:14 PM
I don't think the message "Unable to lookup the cluster by ID" has anything to do with local repositories, that looks like a problem with how you named your cluster. See "clusterName=clusterID=-1". But yes, you had to first create an "ambari.repo" and "hdp.repo" file in /etc/yum.repos.d/ that point to your local repo before you even can install ambari-server and other related HDP packages on that machine. Try starting over. https://community.hortonworks.com/questions/1110/how-to-completely-remove-uninstall-ambari-and-hdp.html
... View more
11-21-2017
04:58 PM
@Michael Bronson m -rf -> This is a Linux/Unix based command which will only delete your Unix/Lrinux based directory created in Unix/Linux file system. Whereas hdfs dfs -rmr /DirectoryPath -> Is for deletion of files/dirs in HDFS filesystem. Incase I miss interpreted your question then and you mean to ask me what is difference between "hdfs dfs -rmr" and "hdfs dfs -rm -rf" then the later one doesn't exist as there is no "-f" parameter to rm command in HDFS filesystem. We only have "-r" as an option for rm command in HDFS to delete the dir and files.
... View more
08-19-2017
05:06 PM
Thanks for the reply! That is exactly my feeling. Because everything is on one page, a search in the Flume docs is easy. Hive is the worst for this... But once you know where you want to get it, it's relatively easy to find it. One question I had was whether when they say "write a hive query", whether that meant that the query had to be one statement? Because I would be able to do it in 2/3 statements but not sure how to do it in only one statement. Of course when you save the query file, it's all executed in one batch but if they only care about the result, I guess it would work either way. Thanks for the info, I guess I'll practice a bit more and sign up for the exam. Cheers
... View more