Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3367 | 05-03-2017 05:13 PM | |
2798 | 05-02-2017 08:38 AM | |
3076 | 05-02-2017 08:13 AM | |
3006 | 04-10-2017 10:51 PM | |
1518 | 03-28-2017 02:27 AM |
12-21-2016
07:58 PM
Simba driver is available as of HDP 2.5 with additional certification from Simba that it works with Kerberos. Here's the official documentation guide for the ODBC driver https://hortonworks.com/wp-content/uploads/2016/08/phoenix-ODBC-guide.pdf
... View more
12-21-2016
01:17 PM
To drive my point home here's more http://blog.mortardata.com/post/60274287605/pig-vs-mapreduce And http://blog.mortardata.com/post/33711299619/8-reasons-you-should-be-using-apache-pig
... View more
12-21-2016
01:06 PM
You can try your use case using Pig and built-in Split function as you'll benefit from underlying query plan optimizations and Tez execution engine compared to pure mapreduce implementation http://pig.apache.org/docs/r0.16.0/basic.html#SPLIT It might be a much more worthwhile investment in your case
... View more
12-21-2016
12:57 PM
1 Kudo
Generally to control output format from reducer you'd use multiple output class https://hadoop.apache.org/docs/r2.7.1/api/org/apache/hadoop/mapreduce/lib/output/MultipleOutputs.html You get best results by writing larger files, not sure what you benefit from splitting a dataset that fits your criteria into smaller chunks, job won't complete until all of the criteria is addressed and in fact I think you'll hurt performance by splitting what is by design a better approach.
... View more
12-21-2016
12:51 PM
If Constantin's awesome answer helped you, please accept the answer to close this thread, otherwise provide your solution or follow up questions for more clarity
... View more
12-21-2016
03:40 AM
1 Kudo
@Edgar Daeds In Apache Zepplin 0.7 there will be an Apache Beam Interpreter that by default will have Java REPL. You can use Beam API to work with Spark, Flink, Mapreduce and Google Dataflow https://zeppelin.apache.org/docs/0.7.0-SNAPSHOT/interpreter/beam.html
... View more
12-20-2016
11:02 PM
Can someone explain the decision to include the following jars in the HDP distribution for HBase? Is this for compatibility? Please provide some technical background on the decision. -rw-r--r-- 1 user user 790250 Nov 15 19:41 netty-3.2.4.Final.jar
-rw-r--r-- 1 user user 1779991 Nov 15 18:11 netty-all-4.0.23.Final.jar
-rw-r--r-- 1 user user 132368 Nov 15 18:12 servlet-api-2.5-6.1.14.jar
-rw-r--r-- 1 user user 105112 Nov 15 18:05 servlet-api-2.5.jar
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Phoenix
12-20-2016
07:46 PM
@Rishit shah can you follow my suggestion and install hive and hcat client tools on the flume nodes? It will dynamically link the jars to proper locations. Just need you to confirm.
... View more