Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3660 | 05-03-2017 05:13 PM | |
| 3017 | 05-02-2017 08:38 AM | |
| 3278 | 05-02-2017 08:13 AM | |
| 3223 | 04-10-2017 10:51 PM | |
| 1686 | 03-28-2017 02:27 AM |
02-13-2016
01:29 PM
1 Kudo
@Rushikesh Deshmukh please provide load statement
... View more
02-12-2016
08:14 PM
sc.parallelize(records).toDF().write.format("orc").save("people")
that method was refactored. There's a new way of writing ORC files. Convert your RDD to DataFrame with toDF() and then write it out as above.Try to use later versions of Spark. http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.4/bk_spark-guide/content/ch_orc-spark.html
... View more
02-12-2016
07:17 PM
@Michel Sumbul the feature is called rquest throttling. It is available in HBase 1.1 and thus in HDP 2.3. More info https://blogs.apache.org/hbase/entry/the_hbase_request_throttling_feature The throttle can then be set from the HBase shell, like so: hbase> set_quota TYPE => THROTTLE, USER => 'uname', LIMIT => '100req/sec' hbase> set_quota TYPE => THROTTLE, TABLE => 'tbl', LIMIT => '10M/sec' hbase> set_quota TYPE => THROTTLE, NAMESPACE => 'ns', LIMIT => 'NONE' our in-house engineer responded in a similar post recently https://community.hortonworks.com/questions/1821/hbase-quota-management.html
... View more
02-12-2016
04:57 PM
@Satish S your issue is all the \ in your query, you use \ outside of quotes, inside the quotes it is interpreted as SQL. That's why your MySQL is complaining. Remove all that and do use '$CONDITIONS', make sure to read the note about wrapping $CONDITIONS with single quotes or double because it does make a difference.
... View more
02-12-2016
01:47 PM
@Saurabh Kumar https://github.com/evidens/json2csv
... View more
02-12-2016
12:30 PM
1 Kudo
@Darpan Patel flume is one directional, you can't push data out of hdfs to fs. We have Apache NiFi that you can leverage for moving logs to and from hdfs.
... View more
02-12-2016
01:08 AM
@bsainiIn the first phase, we have enabled NFSv3 interface access to HDFS. This is done using NFS Gateway, a stateless daemon, that translates NFS protocol to HDFS access protocols as shown in the following diagram. Many instances of such daemon can be run to provide high throughput read/write access to HDFS from multiple clients. As a part of this work, HDFS now has a significant functionality that supports inode ID or file handles, that was done in Apache JIRA HDFS-4489. Source: http://hortonworks.com/blog/simplifying-data-management-nfs-access-to-hdfs/
... View more
02-11-2016
10:36 PM
@PJ Moutrie look at git page last source code check-in is from 7 days ago, project is chugging along. Wait for new version https://git-wip-us.apache.org/repos/asf?p=incubator-atlas.git
... View more
02-11-2016
09:09 PM
@Roberto Sancho https://github.com/twitter/elephant-bird/wiki/Hadoop-2.x-Support So you might have an old jar? Still try built-in jsonloader instead but confirm my answer.
... View more
02-11-2016
09:06 PM
@Roberto Sancho I'm guessing your elephant bird library is compiled for Hadoop 1.x line. See if they offer Hadoop 2 compiled jar. Why not use https://pig.apache.org/docs/r0.15.0/func.html#jsonloadstore ?
... View more