Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3360 | 05-03-2017 05:13 PM | |
2789 | 05-02-2017 08:38 AM | |
3064 | 05-02-2017 08:13 AM | |
3002 | 04-10-2017 10:51 PM | |
1508 | 03-28-2017 02:27 AM |
01-14-2020
11:22 PM
--as-textfile Maps all the columns to their respective datatypes mentioned in –map-column-hive --as-parquet-file Does not change any datatype for the columns mentioned in –map-column-hive Please reply on this, if you have got answer
... View more
01-07-2020
09:16 AM
How did you resolve the issue ?
... View more
12-20-2019
07:16 AM
https://github.com/apache/oozie/blob/9c288fe5cea6f2fbbae76f720b9e215acdd07709/webapp/src/main/webapp/oozie-console.js#L384
... View more
12-19-2019
01:53 PM
@Augustine ,
Thanks for the feedback. I have corrected the answer and thank you for reporting this.
Cheers,
Li
... View more
11-06-2019
10:09 AM
In order to run a balancer on only one data node hdfs balancer -include -f <#datanode name can be specified> this would balance the data load on that particular DN.
... View more
10-11-2019
10:28 PM
It really works after replacing ';' with '\073' ! I met this problem when using split and ';' in crontab hql, it returns " ParseException line 3:121 cannot recognize input near '<EOF>' '<EOF>' '<EOF>' in select expression".
... View more
09-08-2019
06:58 AM
Hi, Can you share the error message from spark history server logs when the spark history UI is crashing? Thanks AKR
... View more
09-02-2019
06:50 PM
@adelgacem can you share sample for how you invoked Oozie using InvokHttp processor? I'm trying to run Oozie workflow from NIFI it will be a great help if you can share NIFI processor configuration details.
... View more
11-09-2018
06:23 AM
@Venu Shanmukappa how did u add the proxy.. can u pls explain
... View more
09-07-2018
01:47 PM
4 Kudos
This is a short how-to leveraging Zeppelin and Solr's native SQL capabilities to query the Ranger audit logs in real time. The capability to query Ranger audits has been in existence for quite a while and there are multiple articles available demonstrating how to apply a Hive external table on top of Ranger audits stored in HDFS. This article demonstrates how to leverage Zeppelin and Solr SQL to query Solr in real time without additional step of creating an external table on top of the HDFS audit. First thing you need is access to your Solr instance. I'm using the default instance packaged with Ambari infra. The Solr admin UI is available at the following address: http://{ambari-infra-ip}:8886. In the UI, you can issue arbitrary queries using standard Solr syntax. I am new to Solr and found the query syntax cumbersome. Instead, I decided to leverage Solr SQL, available as of version 6. HDP 3.0 ships with Solr 7.3. The next step is to set up Zeppelin interpreter for Solr via JDBC. Steps for doing that are available on the Solr website and I'm going to summarize the minimum required configuration for HDP 3. Feel free to copy and modify the properties below: default.driver : org.apache.solr.client.solrj.io.sql.DriverImpl default.url : jdbc:solr://{ambari-infra-ip}:2181/infra-solr?collection=ranger_audits default.user : solr In the artifacts section, add the following entry org.apache.solr:solr-solrj:7.3.1 Be mindful of the port for Zookeeper quorum for the Ranger Solr collection. I found the information browsing the ZK CLI shell. /usr/hdp/current/zookeeper-client/bin/zkCli.sh
[zk: localhost:2181(CONNECTED) 4] ls /infra-solr/collections/ranger_audits So once you enter that information into Zeppelin interpreter, you can now use the %solr command to browse Ranger audits with SQL, just add a new note with Solr interpreter selected. Notice I am using all of the standard fields in Ranger audit, you can find an older version of the schema at the following link. I say older because in HDP 3, Ranger supports multiple clusters and additional fields identifying separate clusters are available, by I digress. The query above will show all current events where result, i.e. access is denied. This is really convenient because you don't need to apply schema and data is available in real time. You can build powerful reporting capabilities on top of what is available in Ranger Admin UI, (in case your question was why even doing that if that info was available already via Ranger). Finally, once you press execute, the data will be shown below: You can now add more plotting libraries and built-in Zeppelin charting capabilities to make very powerful dashboards!
... View more
Labels: