Member since
11-04-2016
87
Posts
9
Kudos Received
3
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
6676 | 11-15-2016 07:16 PM | |
2966 | 11-14-2016 08:05 PM | |
4161 | 11-08-2016 05:00 PM |
03-27-2017
07:30 PM
@Marcy All of them can work. Their access to Hive is commonly done using a Notebook tool called Apache Zeppelin (included in the Hortonworks Data Platform). Hortonworks has many tutorials that can show you step by step on how to connect these: https://hortonworks.com/hadoop-tutorial/using-hive-with-orc-from-apache-spark/ https://hortonworks.com/hadoop-tutorial/getting-started-apache-zeppelin/
... View more
03-14-2017
01:44 PM
I mean yes, i do see this error with single table import... it appears its because of that one column - project_id... Must a table have a key column?
... View more
11-18-2016
11:22 PM
3 Kudos
HDF is best thought of as working with data in motion and HDP as Hadoop, the popular Big Data Platform which in contrast can be seen as data at rest. Both are independent platforms but can are often integrated. When integrated, they are deployed as separate clusters or platforms. Both are open source and Hortonworks provides paid support for each separately. HDF HDF has NiFi, Storm and Kafka (as well as Ambari admin console). These components are used to get data from diverse sources (ranging from social media sites, log files, IoT devices, databases, etc) and send the data to an equally diverse range of target systems. In between, they can transform moving content, make decisions based on moving content, and run analytics on moving content. The actual movement of data is difficult to engineer and these components move data and handle the many challenges in doing so all under the covers with no low-level development needed. See: https://hortonworks.com/products/data-center/hdf/ HDP HDP is more commonly known as the Hadoop or Big Data Platform. It has HDFS, YARN, Map-reduce and Tez processing engines, Hive database, HBase No Sql database, and many other tools to work with Big Data (data in large volumes, wide variety of formats, and fast real-time velocity of arriving on the platform ... the 3 Vs). It stores this data cheaply and flexibly, and uses horizontal scaling of servers to parallel process these 3 Vs of data in a short amount of time (compared to traditional databases which face limits in working with the 3 Vs). What type of processing depends on the out-of-the-box or 3rd party tools used and the use case / business case involved. See: https://hortonworks.com/products/data-center/hdp/ HDF + HDP HDF and HDP are often integrated because HDF is an effective way to get diverse sources of data into HDP to be stored and processed all in one place, to be used by data scientists for example. If this is what you were looking for, let me know by accepting the answer; else, please respond to this answer with further questions and I will follow-up.
... View more
03-17-2017
09:06 PM
any update here? Running into the same issue initctl: Unable to connect to Upstart: Failed to connect to socket /com/ubuntu/upstart: Connection refused
... View more
11-17-2016
07:26 PM
best answer to this is to change the root password in splash screen once you enter Alt+F5 and then use passwd to change the new password for root if a system doesn't ask for new password. And use the new password to login via putty or any client.
... View more
11-08-2016
05:23 PM
Thanks so much jss!!! Wow, you are fast!
... View more
11-10-2016
01:53 PM
@Marcia Hon I wrote an HCC article that walks you through the process of increasing the base size of the CentOS 7 docker vm image. This is the preferred method for making these changes. https://community.hortonworks.com/content/kbentry/65714/how-to-modify-the-default-docker-configuration-on.html
... View more