Member since
08-23-2016
261
Posts
201
Kudos Received
106
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1762 | 01-26-2018 07:28 PM | |
1402 | 11-29-2017 04:02 PM | |
35342 | 11-29-2017 03:56 PM | |
3523 | 11-28-2017 01:01 AM | |
965 | 11-22-2017 04:08 PM |
08-02-2017
03:06 PM
I don't believe that all of the supporting tools are part of the core project, nor would be supported, but you should check with your Hortonworks representative for an official answer.
... View more
08-02-2017
03:02 PM
1 Kudo
If you know how to split it up, you can split the file into others and route each file to its own PutSQL processor. https://community.hortonworks.com/questions/48492/split-data-into-multiple-files-using-nifi-based-on.html
... View more
08-01-2017
08:24 PM
1 Kudo
Hi @Austin Duncan I believe that the PutSQL processor can also do DDL statements such as your CREATE TABLE statement (it just issues the statements). You can also use the ExecuteScript processor and do it via script you write.
... View more
08-01-2017
05:09 PM
Hi @M R I don't believe those components are included in Hortonworks Data Flow out of the box. The release notes in the documentation outline exactly what products (and versions) are supported in HDF. Here is a link for you: https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.0.1/bk_release-notes/content/ch_hdf_relnotes.html
... View more
08-01-2017
03:38 PM
hi @rama Typically, the variable would be defined somewhere earlier in the script that contains the query, or in the CLI, with a SET statement. Something like: SET TB_MASTER=table_name; If you don't see where it was defined, then you would most likely be guessing from the total databases and tables within your system. If the system is very small, that might be feasible. You can use the "show databases" command to list the databases. If you see one that makes sense, you can issue a "use <database_name>" command and then issue a "show tables" to view the list of table names that exist within that Hive database. Here are some links on Hive variables and how they work. https://cwiki.apache.org/confluence/display/Hive/LanguageManual+VariableSubstitution#LanguageManualVariableSubstitution-UsingVariables https://community.hortonworks.com/articles/60309/working-with-variables-in-hive-hive-shell-and-beel.html
... View more
08-01-2017
03:31 PM
2 Kudos
@Dhiraj As far as I am aware, if your local users exist before the LDAP synch, they should still work though Ranger supports only one source. See the following link for a bit more detail: https://community.hortonworks.com/questions/87388/how-to-sync-users-and-groups-from-both-ldap-and-un.html
... View more
08-01-2017
02:38 PM
Hi @Hardik Dave Cluster sizing and planning would require much more detail and in-depth conversation about the use case, the data sizing, etc. A good guide that can help you down the path of sizing your cluster can be found here: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.1/bk_cluster-planning/bk_cluster-planning.pdf I would suggest using Ambari to manage your cluster where the memory allocation and settings/config will be much more visible in the UI for each technology/service being used in your cluster.
... View more
07-31-2017
04:14 PM
Hi @Vikas Malviya the default is usually admin/admin but if that is not working for you, can you ssh as a super user and run 'ambari-admin-password-reset'
... View more
07-31-2017
04:09 PM
hi @Manoj Dixit We have a case open on the issue intenrally now 🙂
... View more
07-31-2017
03:50 PM
hi @Sahil Jindal Please check your first post on this topic.
... View more