Member since
12-09-2015
115
Posts
43
Kudos Received
12
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
6339 | 07-10-2017 09:38 PM | |
3917 | 04-10-2017 03:24 PM | |
670 | 03-04-2017 04:08 PM | |
2325 | 02-17-2017 10:42 PM | |
3430 | 02-17-2017 10:41 PM |
02-20-2017
07:46 PM
beeline> !connect jdbc:hive2://master1.chrsv.com:8443/default;transportMode=http;httpPath=gateway/default/hive;ssl=false
Connecting to jdbc:hive2://master1.chrsv.com:8443/default;transportMode=http;httpPath=gateway/default/hive;ssl=false
Enter username for jdbc:hive2://master1.chrsv.com:8443/default;transportMode=http;httpPath=gateway/default/hive;ssl=false: rvchinta
Enter password for jdbc:hive2://master1.chrsv.com:8443/default;transportMode=http;httpPath=gateway/default/hive;ssl=false: ********
Connected to: Apache Hive (version 1.2.1000.2.5.3.0-37)
Driver: Hive JDBC (version 1.2.1000.2.5.3.0-37)
Transaction isolation: TRANSACTION_REPEATABLE_READ
0: jdbc:hive2://master1.chrsv.com:8443/defaul> show databases;
+----------------+--+
| database_name |
+----------------+--+
| default |
| test |
+----------------+--+ 2017-02-20 14:43:45,115 INFO [HiveServer2-HttpHandler-Pool: Thread-57]: thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(398)) - Failed to authenticate with http/_HOST kerberos principal, trying with hive/_HOST kerberos principal
2017-02-20 14:43:45,116 ERROR [HiveServer2-HttpHandler-Pool: Thread-57]: thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(406)) - Failed to authenticate with hive/_HOST kerberos principal
2017-02-20 14:43:45,116 ERROR [HiveServer2-HttpHandler-Pool: Thread-57]: thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(209)) - Error:
org.apache.hive.service.auth.HttpAuthenticationException: java.lang.reflect.UndeclaredThrowableException
at org.apache.hive.service.cli.thrift.ThriftHttpServlet.doKerberosAuth(ThriftHttpServlet.java:407)
at org.apache.hive.service.cli.thrift.ThriftHttpServlet.doPost(ThriftHttpServlet.java:159)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:565)
at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:479)
at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:225)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1031)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:406)
at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:186)
at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:965)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
at org.eclipse.jetty.server.Server.handle(Server.java:349)
at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:449)
at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.content(AbstractHttpConnection.java:925)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:952)
at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
at org.eclipse.jetty.server.AsyncHttpConnection.handle(AsyncHttpConnection.java:76)
at org.eclipse.jetty.io.nio.SelectChannelEndPoint.handle(SelectChannelEndPoint.java:609)
at org.eclipse.jetty.io.nio.SelectChannelEndPoint$1.run(SelectChannelEndPoint.java:45)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.UndeclaredThrowableException
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1742)
at org.apache.hive.service.cli.thrift.ThriftHttpServlet.doKerberosAuth(ThriftHttpServlet.java:404)
... 23 more
Caused by: org.apache.hive.service.auth.HttpAuthenticationException: Authorization header received from the client is empty.
at org.apache.hive.service.cli.thrift.ThriftHttpServlet.getAuthHeader(ThriftHttpServlet.java:548)
at org.apache.hive.service.cli.thrift.ThriftHttpServlet.access$100(ThriftHttpServlet.java:74)
at org.apache.hive.service.cli.thrift.ThriftHttpServlet$HttpKerberosServerAction.run(ThriftHttpServlet.java:449)
at org.apache.hive.service.cli.thrift.ThriftHttpServlet$HttpKerberosServerAction.run(ThriftHttpServlet.java:412)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724)
... 24 more
2017-02-20 14:43:45,172 INFO [HiveServer2-HttpHandler-Pool: Thread-57]: thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(145)) - Could not validate cookie sent, will try to generate a new cookie
2017-02-20 14:43:45,176 INFO [HiveServer2-HttpHandler-Pool: Thread-57]: thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(204)) - Cookie added for clientUserName knox
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Knox
02-17-2017
10:42 PM
this is because mysql is external to ambari and when kerberos is enabled ambari is not smart enough to recognize mysql and it didnot create keytabs for mysql. that was the reason hive was not able to start. i still need to find out a way to create keytabs for non ambari components. as of now i moved these components to another server where all the services were deployed through ambari. thanks to all for your help so far.
... View more
02-17-2017
10:41 PM
this is because mysql is external to ambari and when kerberos is enabled ambari is not smart enough to recognize mysql and it didnot create keytabs for mysql. that was the reason hive was not able to start. i still need to find out a way to create keytabs for non ambari components. as of now i moved these components to another server where all the services were deployed through ambari. thanks to all for your help so far.
... View more
02-14-2017
03:40 PM
I have internal mysql (ambari is not aware of it), this mysql will be used for oozie service. when we enable kerberos. will ambari create keytab entries for mysql? if not how do we create it manually?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Oozie
-
Kerberos
02-08-2017
07:54 PM
@Ameet Paranjape we are talking about hive metastore....hiveserver itself is not talking to hivemetastore...so connecting to beeline does not help.
... View more
02-08-2017
07:51 PM
Any suggestions/script to kill yarn application if it is running more than 20 minutes?
... View more
Labels:
- Labels:
-
Apache YARN
02-03-2017
09:37 PM
Feb 03 15:23:55 master2.chrsv.com krb5kdc[3363](info): TGS_REQ (4 etypes {18 17 16 23}) 192.168.56.21: ISSUE: authtime 1486153210, etypes {rep=18 tkt=18 ses=18}, nn/master1.chrsv.com@KERBEROS.COM for nn/master1.chrsv.com@KERBEROS.COM
Feb 03 15:24:00 master2.chrsv.com krb5kdc[3363](info): AS_REQ (4 etypes {18 17 16 23}) 192.168.56.63: ISSUE: authtime 1486153440, etypes {rep=18 tkt=18 ses=18}, hdfs-hdp@KERBEROS.COM for krbtgt/KERBEROS.COM@KERBEROS.COM
Feb 03 15:24:08 master2.chrsv.com krb5kdc[3363](info): AS_REQ (4 etypes {18 17 16 23}) 192.168.56.22: ISSUE: authtime 1486153448, etypes {rep=18 tkt=18 ses=18}, ambari-qa-hdp@KERBEROS.COM for krbtgt/KERBEROS.COM@KERBEROS.COM
Feb 03 15:24:12 master2.chrsv.com krb5kdc[3363](info): TGS_REQ (4 etypes {18 17 16 23}) 192.168.56.22: ISSUE: authtime 1486153222, etypes {rep=18 tkt=18 ses=18}, nn/master2.chrsv.com@KERBEROS.COM for HTTP/master2.chrsv.com@KERBEROS.COM I see services but not hive in krb5.log
... View more
02-03-2017
09:32 PM
@icocio [hive@master2 ~]$ klist -kte /etc/security/keytabs/hive.service.keytab
Keytab name: FILE:/etc/security/keytabs/hive.service.keytab
KVNO Timestamp Principal
---- ----------------- --------------------------------------------------------
2 02/03/17 15:12:29 hive/master2.chrsv.com@KERBEROS.COM (arcfour-hmac)
2 02/03/17 15:12:29 hive/master2.chrsv.com@KERBEROS.COM (des-cbc-md5)
2 02/03/17 15:12:29 hive/master2.chrsv.com@KERBEROS.COM (aes256-cts-hmac-sha1-96)
2 02/03/17 15:12:29 hive/master2.chrsv.com@KERBEROS.COM (aes128-cts-hmac-sha1-96)
2 02/03/17 15:12:29 hive/master2.chrsv.com@KERBEROS.COM (des3-cbc-sha1)
[hive@master2 ~]$
... View more
02-03-2017
09:05 PM
not able to start metastore....it is kerborized using Ambari, version 2.5.3 [hive@master2 ~]$ klist
klist: No credentials cache found (ticket cache FILE:/tmp/krb5cc_501)
[hive@master2 ~]$ kinit -k -t /etc/security/keytabs/hive.service.keytab hive/master2.chrsv.com@KERBEROS.COM
[hive@master2 ~]$ klist Ticket cache: FILE:/tmp/krb5cc_501 Default principal: hive/master2.chrsv.com@KERBEROS.COM
Valid starting Expires Service principal 02/03/17 14:55:41 02/04/17 14:55:41 krbtgt/KERBEROS.COM@KERBEROS.COM renew until 02/03/17 14:55:41 [hive@master2 ~]$
... View more
Labels:
- Labels:
-
Apache Hive
02-03-2017
08:56 PM
my bad..i didnot check syntax earlier... [hive@master2 ~]$ kinit -k -t /etc/security/keytabs/hive.service.keytab hive/master2.chrsv.com@KERBEROS.COM
[hive@master2 ~]$ klist
Ticket cache: FILE:/tmp/krb5cc_501
Default principal: hive/master2.chrsv.com@KERBEROS.COM
Valid starting Expires Service principal
02/03/17 14:55:41 02/04/17 14:55:41 krbtgt/KERBEROS.COM@KERBEROS.COM
renew until 02/03/17 14:55:41
[hive@master2 ~]$
... View more
02-03-2017
08:15 PM
@Ameet Paranjape [hive@master2 ~]$ kinit -k -t /etc/security/keytabs/hive.service.keytab kinit: Cannot determine realm for host (principal host/master2.chrsv.com@) Not sure why it is not picking since all these were setup by Ambari...howevr when i do kadmin i can see the principle as hive/master1.chrsv.com@KERBEROS.COM
hive/master2.chrsv.com@KERBEROS.COM
hive/worker1.chrsv.com@KERBEROS.COM
hive/worker2.chrsv.com@KERBEROS.COM
... View more
02-03-2017
05:15 PM
I got Mysql which is metadata for oozie,hive,amabri. this was setup before cluster is setup. I do not see mysql as a service in hive home. I've enabled local MIT KDC and i see below in metastore.log...i donot see it anywhere it is being authenticated to kdc 2017-02-03 11:00:13,343 INFO [main]: timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(82)) - Initializing Timeline metrics sink.
2017-02-03 11:00:13,345 INFO [main]: timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(100)) - Identified hostname = master2.chrsv.com, serviceName = hivemetastore
2017-02-03 11:00:14,257 INFO [main]: timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(118)) - Collector Uri: http://worker1.chrsv.com:6188/ws/v1/timeline/metrics
2017-02-03 11:00:14,592 INFO [main]: impl.MetricsSinkAdapter (MetricsSinkAdapter.java:start(206)) - Sink timeline started
2017-02-03 11:00:15,133 INFO [main]: impl.MetricsSystemImpl (MetricsSystemImpl.java:startTimer(376)) - Scheduled snapshot period at 10 second(s).
2017-02-03 11:00:15,133 INFO [main]: impl.MetricsSystemImpl (MetricsSystemImpl.java:start(192)) - hivemetastore metrics system started
2017-02-03 11:00:15,938 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:newRawStore(667)) - 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
2017-02-03 11:00:16,495 INFO [main]: metastore.ObjectStore (ObjectStore.java:initializeHelper(370)) - ObjectStore, initialize called
2017-02-03 11:00:26,552 INFO [main]: metastore.ObjectStore (ObjectStore.java:getPMF(474)) - Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes="Table,Database,Type,FieldSchema,Order"
2017-02-03 11:00:39,897 INFO [main]: metastore.MetaStoreDirectSql (MetaStoreDirectSql.java:<init>(138)) - Using direct SQL, underlying DB is MYSQL
2017-02-03 11:00:39,915 INFO [main]: metastore.ObjectStore (ObjectStore.java:setConf(284)) - Initialized ObjectStore
2017-02-03 11:00:41,013 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:createDefaultRoles_core(741)) - Added admin role in metastore
2017-02-03 11:00:41,034 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:createDefaultRoles_core(750)) - Added public role in metastore
2017-02-03 11:00:41,131 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:addAdminUsers_core(790)) - No user is added in admin role, since config is empty
2017-02-03 11:00:41,139 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:init(525)) - Begin calculating metadata count metrics.
2017-02-03 11:00:41,233 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:init(527)) - Finished metadata count metrics: 1 databases, 0 tables, 0 partitions.
2017-02-03 11:00:42,847 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:startMetaStore(6298)) - Starting DB backed MetaStore Server with SetUGI enabled
2017-02-03 11:00:42,861 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:startMetaStore(6352)) - Started the new metaserver on port [9083]...
2017-02-03 11:00:42,862 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:startMetaStore(6354)) - Options.minWorkerThreads = 200
2017-02-03 11:00:42,862 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:startMetaStore(6356)) - Options.maxWorkerThreads = 100000
2017-02-03 11:00:42,862 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:startMetaStore(6358)) - TCP keepalive = true
... View more
Labels:
- Labels:
-
Apache Hive
02-02-2017
06:10 PM
@SBandaru @Prabhu M question guys-- what is the best sequence to make env Kerborized Local KDC for SP & AD for UP with no SSL. As of now i got ENV up and running with no kerborisation.
... View more
01-30-2017
12:30 PM
@Prabhu M thank you very much for your response. let me go through suggested links, it might take a while to read/understand/implement.
... View more
01-29-2017
06:28 PM
2 Kudos
I got HDP 2.5.3 installed. HA/Local MIT KDC enabled. I want to integrate it with AD using sssd. any good documentation available?
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
01-29-2017
04:35 PM
@Tibor Kiss thank you for your response. indeed it is kerberoised env. I've enabled it through ambari. All the above parameters were set by ambari when i enabled kerberos. still it fails with error reported earlier.
... View more
01-28-2017
09:21 PM
[root@master2 ~]# hive
Logging initialized using configuration in file:/etc/hive/2.5.3.0-37/0/hive-log4j.properties
Exception in thread "main" java.lang.RuntimeException: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS]
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:543)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:233)
at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
Caused by: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2160)
at org.apache.hadoop.hdfs.DistributedFileSystem$25.doCall(DistributedFileSystem.java:1423)
at org.apache.hadoop.hdfs.DistributedFileSystem$25.doCall(DistributedFileSystem.java:1419)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1419)
at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1443)
at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:628)
at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:576)
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:529)
... 8 more
Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS]
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1552)
at org.apache.hadoop.ipc.Client.call(Client.java:1496)
at org.apache.hadoop.ipc.Client.call(Client.java:1396)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
at com.sun.proxy.$Proxy12.getFileInfo(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:816)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:278)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:194)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:176)
at com.sun.proxy.$Proxy13.getFileInfo(Unknown Source)
at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2158)
... 16 more
[root@master2 ~]#
... View more
01-28-2017
08:40 PM
HDP 2.5.3 / Kerberos / hiemetastore db is mysql 2017-01-28 13:16:01,085 WARN [main]: server.HiveServer2 (HiveServer2.java:startHiveServer2(508)) - Error starting HiveServer2 on attempt 2, will retry in 60 seconds
java.lang.RuntimeException: Error applying authorization policy on hive configuration: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS]
at org.apache.hive.service.cli.CLIService.init(CLIService.java:114)
at org.apache.hive.service.CompositeService.init(CompositeService.java:59)
at org.apache.hive.service.server.HiveServer2.init(HiveServer2.java:122)
at org.apache.hive.service.server.HiveServer2.startHiveServer2(HiveServer2.java:474)
at org.apache.hive.service.server.HiveServer2.access$700(HiveServer2.java:87)
at org.apache.hive.service.server.HiveServer2$StartOptionExecutor.execute(HiveServer2.java:720)
at org.apache.hive.service.server.HiveServer2.main(HiveServer2.java:593)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:233)
at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
Caused by: java.lang.RuntimeException: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS]
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:543)
at org.apache.hive.service.cli.CLIService.applyAuthorizationConfigPolicy(CLIService.java:127)
at org.apache.hive.service.cli.CLIService.init(CLIService.java:112)
... View more
Labels:
12-02-2016
02:17 PM
1 Kudo
I was able to create hive table on top of json files. below is the syntax i used to create external table..so i donot have to move data, all i need is to add partition CREATE EXTERNAL
TABLE hdfs_audit( access string, agenthost string, cliip string, enforcer string, event_count bigint, event_dur_ms bigint, evttime timestamp, id string, logtype string, policy bigint, reason string, repo string, repotype bigint, requser string, restype string, resource string, result bigint, seq_num bigint) PARTITIONED BY ( evt_time string) ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe' STORED AS
INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' OUTPUTFORMAT
'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' LOCATION 'hdfs://CLUSTERNAME/ranger/database/hdfs'; Add partition: ALTER
TABLE ranger_audit.hdfs_audit ADD PARTITION (evt_time='20160601') LOCATION
'/ranger/audit/hdfs/20160601/hdfs/20160601';
... View more
11-10-2016
12:59 AM
I need some help to create hive table for the below format. this is ranger audit info, will create partitions accordingly. below info is just one line, {"repoType":1,"repo":"abc_hadoop","reqUser":"ams","evtTime":"2016-09-19 13:14:40.197","access":"READ","resource":"/ambari-metrics-collector/hbase/data/hbase/meta/1588230740/info/ed3e52d8b86e4800801539fc4a7b1318","resType":"path","result":1,"policy":41,"reason":"/ambari-metrics-collector/hbase/data/hbase/meta/1588230740/info/ed3e52d8b86e4800801539fc4a7b1318","enforcer":"ranger-acl","cliIP":"123.129.390.140","agentHost":"hostname.sample.com","logType":"RangerAudit","id":"94143368-600c-44b9-a0c8-d906b4367537","seq_num":1240883,"event_count":1,"event_dur_ms":0}
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Ranger
09-21-2016
12:35 PM
@Michael Young thank you for your response, but i do not want to save my password in plain text even for a short period of time. thanks, Raj
... View more
09-16-2016
05:39 PM
2 Kudos
goal is to connect through beeline and pass password as a variable...it just works fine, but we can see the password in plain text if we do ps -ef|grep hive
[root@sandbox ~]# beeline -u 'jdbc:hive2://sandbox.hortonworks.com:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2' -n rajesh -w pass
WARNING: Use "yarn jar" to launch YARN applications.
Connecting to jdbc:hive2://sandbox.hortonworks.com:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2
Connected to: Apache Hive (version 1.2.1000.2.5.0.0-817)
Driver: Hive JDBC (version 1.2.1000.2.5.0.0-817)
Transaction isolation: TRANSACTION_REPEATABLE_READ
Beeline version 1.2.1000.2.5.0.0-817 by Apache Hive
0: jdbc:hive2://sandbox.hortonworks.com:2181/>
... View more
Labels:
- Labels:
-
Apache Hive
08-31-2016
01:05 PM
save password in text file and connect as below:) [root@sandbox ~]# beeline -u 'jdbc:hive2://sandbox.hortonworks.com:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2' -n rajesh -w pass WARNING: Use "yarn jar" to launch YARN applications. Connecting to jdbc:hive2://sandbox.hortonworks.com:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2 Connected to: Apache Hive (version 1.2.1000.2.5.0.0-817) Driver: Hive JDBC (version 1.2.1000.2.5.0.0-817) Transaction isolation: TRANSACTION_REPEATABLE_READ Beeline version 1.2.1000.2.5.0.0-817 by Apache Hive 0: jdbc:hive2://sandbox.hortonworks.com:2181/>
... View more
08-25-2016
06:34 PM
I'm planing to access Hbase using Phoenix. I'm able to access it with no security. i mean it is not authenticating. I want the jdbc connection go though knox so it will authenticate against AD. current url is jdbc:phoenix:abcdef.abc.com:2181:/hbase-unsecure
... View more
Labels:
- Labels:
-
Apache Knox
-
Apache Phoenix
07-27-2016
02:10 PM
@Kuldeep Kulkarni I got the same error message. only difference is my env is kerborized & both my rm's are not in standby mode. [yarn@m1 root]$ yarn rmadmin -getServiceState rm1 standby [yarn@m1 root]$ yarn rmadmin -getServiceState rm2 active Ambari doesn't show the state of RM, but getting same exception as above. i tried to switch the roles and that didnot help. Any help is appreciated.
... View more
07-20-2016
04:52 PM
thank you, this gives me a good idea. let me play with ranger and see what i can accomplish
... View more
04-04-2016
03:48 PM
desc function reflect;
Error: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.parse.SemanticException: UDF reflect is not allowed (state=08S01,code=1) it is working with hive though: hive> describe function reflect;
OK
reflect(class,method[,arg1[,arg2..]]) calls method with reflection
Time taken: 1.973 seconds, Fetched: 1 row(s)
... View more
04-04-2016
03:40 PM
2 Kudos
not able to use refect function using beeline, same query just works fine with hivecli. both in hivecli & beeline when i search for show functions i could see reflect .. I donot see any denials in ranger SELECT reflect("java.net.URLDecoder","decode",search_query) FROM jlp_endeca_searches limit 5;
Error: Error while compiling statement: FAILED: SemanticException UDF reflect is not allowed (state=42000,code=40000)
... View more
Labels:
- Labels:
-
Apache Hive
03-18-2016
09:10 PM
3 Kudos
it is an issue with meta data..we just dropped it and recreated the table and all is well...
... View more