Member since
09-23-2015
800
Posts
898
Kudos Received
185
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 7358 | 08-12-2016 01:02 PM | |
| 2708 | 08-08-2016 10:00 AM | |
| 3674 | 08-03-2016 04:44 PM | |
| 7214 | 08-03-2016 02:53 PM | |
| 1864 | 08-01-2016 02:38 PM |
03-14-2016
08:16 PM
2 Kudos
Any idea why this parameter is required at all? Hiveserver should know which principal it is started with. Why would you have to tell it again?
... View more
03-13-2016
09:50 PM
3 Kudos
There is actually a way to change the numbers of mappers in Pig. Pig uses a CombineFileInputFormat to merge small files into bigger map tasks. This is enabled by default and can be modified with the following parameters: For the rest what Artem said.
pig.maxCombinedSplitSize – Specifies the size, in bytes, of data to be processed by a single map. Smaller files are combined untill this size is reached. pig.splitCombination – Turns combine split files on or off (set to “true” by default).
... View more
03-11-2016
09:21 AM
3 Kudos
Had that problem before. I didn't find any great websites around it. However the source code of the piggybank functions contains some really good documentation in the javadocs. https://pig.apache.org/docs/r0.8.1/api/org/apache/pig/piggybank/storage/MultiStorage.html or directly the source code, many of the functions are pretty straight forward to understand from code: http://svn.apache.org/viewvc/pig/trunk/contrib/piggybank/java/src/main/java/org/apache/pig/piggybank/storage/MultiStorage.java?view=co I didn't find anything better, doesn't mean that it doesn't exist.
... View more
03-10-2016
09:48 AM
1 Kudo
Ambari the management tool for HDP ( Hue is the working environment. ) Ambari can normaly found under port 8080 or 8443 on the ambari server ( if you use a sandbox for example )
... View more
03-09-2016
05:40 PM
2 Kudos
Its not a config, every hiveserver2 gets the same config. Go to Summary on the hive tab and you will see a list of all hiveservers metastore services etc. Click on the hiveserver2 one and you see the host its running on.
... View more
03-09-2016
04:31 PM
1 Kudo
Not really, its more a server but since its just a proxy similarly to the httpfs servers they didn't add it as the own service. After all Phoenix is always installed. ( I normally try to avoid the Query server anyway for performance issues. Diretly accessing Phoenix over the Hbase API is normally the better choice as long as your clients have access to all Regionservers )
... View more
03-09-2016
03:15 PM
2 Kudos
Phoenix is a library always added to HBase. There are also Phoenix Query servers which are essentially Phoenix proxies for clients who cannot access all data nodes. They can be added to a host on the host page Click on a host, press +Add: Phoenix Query Server ( more like a client than a service since Phoenix is already part of HBAse )
... View more
03-09-2016
11:23 AM
2 Kudos
Are you sure that uit works. I.e. have you restarted the Hive services after changing authentication to LDAP. I am pretty sure it should block you after. I did PAM before a couple of times and I assure you when I enable that nothing works anymore if any of the other dependencies are not setup correctly. https://community.hortonworks.com/articles/591/using-hive-with-pam-authentication.html
... View more
03-09-2016
11:20 AM
3 Kudos
I wrote a little article on getting started with SQL STD Auth. I implemented it and did quite like it. https://community.hortonworks.com/content/kbentry/597/getting-started-with-sqlstdauth.html
... View more
03-08-2016
05:02 PM
3 Kudos
By itself any HDFS user can use Hive. If you do not have kerberos: 1) Just add the user to Linux with useradd ( or add to your ldap ) 2) create user home directory as user hdfs do hadoop fs -mkdir /home/myuser hadoop fs-chown myuser /home/myuser And that's it you should now be able to run queries as that user. Now there is a second part for secure clusters If you have kerberos: 1) Add user to your KDC 2) kinit as the user 3) run hive Using hiveserver2 with beeline or jdbc driver: Depends on which security you have configured for hiveserver2 - None, just specify myuser as user ( -n myuser ) - PAM uses Linux so give -n myuser -p myuserlinuxpassword - LDAP uses an LDAP server lets assume its the same as your linux user - kerberos needs a kinit and specify the principal in the jdbc url Finally your user needs access to the tables: No authorization - make sure the tables you want to read are readable by that user or enable doAs=false Ranger - Add access to the table in Ranger portal SQLStdAuth - Grant access to the table using GRANT command
... View more