Member since
09-18-2015
3274
Posts
1159
Kudos Received
426
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2557 | 11-01-2016 05:43 PM | |
| 8474 | 11-01-2016 05:36 PM | |
| 4848 | 07-01-2016 03:20 PM | |
| 8166 | 05-25-2016 11:36 AM | |
| 4311 | 05-24-2016 05:27 PM |
04-15-2016
05:58 PM
As far as I understand, tblproperties is just a list of key/value pairs. You can specify whatever you want on it (even if it's not predefined, e.g. 'weather'='warm'). But of course there must be code logic that interprets your tblproperties and do something, otherwise it's useless.
... View more
04-15-2016
09:09 AM
4 Kudos
Original Post
Calcite is a highly customizable engine for parsing and planning queries on data in a wide variety of formats. It allows database-like access, and in particular a SQL interface and advanced query optimization, for datanot residing in a traditional database.
Apache Calcite is a dynamic data management framework.
It contains many of the pieces that comprise a typical database management system, but omits some key functions: storage of data, algorithms to process data, and a repository for storing metadata.
Calcite intentionally stays out of the business of storing and processing data. As we shall see, this makes it an excellent choice for mediating between applications and one or more data storage locations and data processing engines. It is also a perfect foundation for building a database: just add data. Source
Tutorial https://calcite.apache.org/docs/tutorial.html
Demo:
Read DEPT and EMPS table
Create a test table based on existing csv example. Read the tutorial link to understand the model.json and schema.
In the demo, you can see that I am running explain plan on the queries and then I used smart.json to change the plan.
Watch the demo and then read the following links
model.json https://calcite.apache.org/docs/tutorial.html#schema-discovery
Query tuning https://calcite.apache.org/docs/tutorial.html#optimizing-queries-using-planner-rules
Calcite https://calcite.apache.org/
This page describes the SQL dialect recognized by Calcite’s default SQL parser.
Adapters
JDBC driver
Calcite is embedded in Drill, Hive and Kylin.
... View more
Labels:
03-28-2016
09:29 PM
2 Kudos
It appears this was a bug that was filed with MySQL pertaining to the JDBC driver version I was using. I was originally using mysql-connector-java-5.1.17.jar. I have since upgraded to mysql-connector-java-5.1.38-bin.jar and this issue has gone away.
... View more
03-08-2017
02:35 PM
Hi @Andrew Grande and @Neeraj Sabharwal 1) I was able to use the TransFormXml processor to convert the xml in the following format <?xml version="1.0" encoding="UTF-8"?> -<ROWSET> <ROW kind="element" name="Product" pid="" id="d1e1"> </ROW> <ROW kind="attribute" name="Type" pid="d1e1" id="d1e1a1373">Laptop</ROW> <ROW kind="element" name="Notebook" pid="d1e1" id="d1e3"> </ROW> <ROW kind="attribute" name="Brand" pid="d1e3" id="d1e3a1403">HP</ROW> <ROW kind="attribute" name="Model" pid="d1e3" id="d1e3a1938">Pavilion dv6-3132TX Notebook</ROW> <ROW kind="element" name="Harddisk" pid="d1e3" id="d1e5">640 GB</ROW> <ROW kind="element" name="Processor" pid="d1e3" id="d1e8">Intel Core i7</ROW> <ROW kind="element" name="RAM" pid="d1e3" id="d1e11">4 GB</ROW> <ROW kind="element" name="Notebook" pid="d1e1" id="d1e15"> </ROW> <ROW kind="attribute" name="Brand" pid="d1e15" id="d1e15a1403">HP</ROW> <ROW kind="attribute" name="Model" pid="d1e15" id="d1e15a1938">HP Pavilion dv6-3032TX Notebook</ROW> <ROW kind="element" name="Harddisk" pid="d1e15" id="d1e17">640 GB</ROW> <ROW kind="element" name="Processor" pid="d1e15" id="d1e20">Intel Core i7</ROW> <ROW kind="element" name="RAM" pid="d1e15" id="d1e23">6 GB</ROW> <ROW kind="element" name="Notebook" pid="d1e1" id="d1e27"> </ROW> <ROW kind="attribute" name="Brand" pid="d1e27" id="d1e27a1403">Toshiba</ROW> <ROW kind="attribute" name="Model" pid="d1e27" id="d1e27a1938">Satellite A660/07R 3D Notebook</ROW> <ROW kind="element" name="Harddisk" pid="d1e27" id="d1e29">640 GB</ROW> <ROW kind="element" name="Processor" pid="d1e27" id="d1e32">Intel Core i7</ROW> <ROW kind="element" name="RAM" pid="d1e27" id="d1e35">4 GB</ROW> <ROW kind="element" name="Notebook" pid="d1e1" id="d1e39"> </ROW> <ROW kind="attribute" name="Brand" pid="d1e39" id="d1e39a1403">Toshiba</ROW> <ROW kind="attribute" name="Model" pid="d1e39" id="d1e39a1938">Satellite A660/15J Notebook</ROW> <ROW kind="element" name="Harddisk" pid="d1e39" id="d1e41">640 GB</ROW> <ROW kind="element" name="Processor" pid="d1e39" id="d1e44">Intel Core i5</ROW> <ROW kind="element" name="RAM" pid="d1e39" id="d1e47">6 GB</ROW> </ROWSET> 2) my question is how to convert the following xquery in the sample for Nifi's EvaluateXQuery : (In the example, oracle is using.) SELECT x.* FROM xml_test t
, XMLTable('/ROWSET/ROW' passing xmltransform(t.object_value, xmltype(:xsldoc)) columns node_id varchar2(100) path '@id'
, node_name varchar2(30) path '@name'
, node_value varchar2(2000) path 'text()'
, parent_node_id varchar2(100) path '@pid'
, node_kind varchar2(30) path '@kind' ) x
;
... View more
06-01-2016
08:58 AM
Hi @Sree Venkata. To ad to Neeraj's already excellent answer and to follow your comment, NiFi now *does* support kerberised clusters. Also there is now an RDBMS connector, although I'd still say, use SQOOP if you're transferring very large chunks of RDBMS data and you want it parellised across the whole hadoop cluster, use NiFi if youve got smaller chunks to transfer that can be parallelised over a smaller NiFi cluster. Hope that (in combination with Neeraj's answer) fulfills your requirements.
... View more
03-15-2016
06:33 PM
1 Kudo
I have installed only zookeeper.
... View more
03-16-2016
03:55 AM
1 Kudo
@Jitendra Yadav - I have tested and it's working! Thank you. https://community.hortonworks.com/questions/23151/how-to-specify-multiple-fencing-methods-in-dfshafe.html#answer-23221
... View more
10-06-2016
08:47 PM
1 Kudo
How to remove clients from the worker nodes? Thanks
... View more
05-05-2016
06:06 PM
1 Kudo
We had similar issues with the hive interpreter while trying to run aggregations and gouping by columns: 1. Hive interpreter cannot be declared directly in notebook by using %hive. Interpreter must already be set to hive 2. First line in editor must be blank and Hive QL statements must start on second line, otherwise a NullPointer exception will be thrown after you submit job. This threw us off. Somehow we started the statement on the second line and it executed without errors. Then, when we went back and put the statement on the first line, it failed again. Moved the statement to the second line, with the first line blank, and it executed without any errors. Strange.
... View more
03-23-2017
02:28 AM
Look for "Hortonworks Sandbox Archive" under https://hortonworks.com/downloads/#sandbox and click on "Expand" to find older versions of sandbox.
... View more