Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3649 | 05-03-2017 05:13 PM | |
| 3007 | 05-02-2017 08:38 AM | |
| 3262 | 05-02-2017 08:13 AM | |
| 3216 | 04-10-2017 10:51 PM | |
| 1680 | 03-28-2017 02:27 AM |
03-11-2016
06:37 PM
@Sarat Pidaparthi please look in /usr/hdp/current/pig-client/lib directory or /usr/hdp/version/pig/lib. Run pig service check, you need not set path for PIG_HOME as Ambari will take care of that for you. Do you have pig client installed on your edge node?
... View more
03-11-2016
03:11 PM
@ARUNKUMAR RAMASAMY HBase has a concept of filters. For your gets and scans, you can setup a filter for your specific need. Consider using the following filter and comparator classes for case-incensitive queries. http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/filter/SubstringComparator.html and this is the class for filter https://hbase.apache.org/book.html#client.filter.cv.scvf You can also try RegexStringComparator http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/filter/RegexStringComparator.html A SingleColumnValueFilter (see: http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html) can be used to test column values for equivalence ( CompareOp.EQUAL ), inequality ( CompareOp.NOT_EQUAL ), or ranges (e.g., CompareOp.GREATER ). The following is an example of testing equivalence of a column to a String value "my value"… SingleColumnValueFilter filter = new SingleColumnValueFilter(
cf,
column,
CompareOp.EQUAL,
Bytes.toBytes("my value")
);
scan.setFilter(filter);
// example with Case Incensitive substring comparator
SingleColumnValueFilter scvf =
new SingleColumnValueFilter("col", CompareOp.EQUAL,
new SubstringComparator("substr"));
Cloudera has good documentation for HBase filtering in Java and CLI http://www.cloudera.com/documentation/enterprise/5-2-x/topics/admin_hbase_filtering.html
... View more
03-11-2016
02:12 PM
@Sunile Manjee I guess I can't avoid readming docs today :), thank you.
... View more
03-11-2016
01:51 PM
1 Kudo
Should Ranger and Ranger KMS be on the same host or separate? I wonder about any security implications, DR, anything else? Please provide pros/cons.
... View more
Labels:
- Labels:
-
Apache Ranger
03-11-2016
11:54 AM
1 Kudo
Did you register the jar? Please confirm and I'll test it.
... View more
03-11-2016
11:52 AM
Similar question has been asked before https://community.hortonworks.com/questions/20487/store-output-file-as-3-files-using-pig.html I am going to repeat my findings over @Rich Raposa answer. It's only relevant if it's not for the purposes of the exam. This question was bothering me and I needed to try it out. here's a full script, piggybank is both in pig-client/lib and in pig-client directory REGISTER /usr/hdp/current/pig-client/piggybank.jar;
A = LOAD 'data2' USING PigStorage() as (url, count);
fs -rm -R output;
STORE A INTO 'output' USING org.apache.pig.piggybank.storage.MultiStorage('output', '0'); dataset is 1
2
3
4
5 output is -rw-r--r-- 3 root hdfs 3 2016-03-18 01:51 /user/root/output/1/1-0,000
Found 1 items
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:51 /user/root/output/2/2-0,000
Found 1 items
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:51 /user/root/output/3/3-0,000
Found 1 items
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:51 /user/root/output/4/4-0,000
Found 1 items
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:51 /user/root/output/5/5-0,000
-rw-r--r-- 3 root hdfs 0 2016-03-18 01:51 /user/root/output/_SUCCESS
each file has one line [root@sandbox ~]# hdfs dfs -cat /user/root/output/5/5-0,000
5 in case of @Rich Raposa example the output directory would look like so: [root@sandbox ~]# hdfs dfs -ls output3
Found 6 items
-rw-r--r-- 3 root hdfs 0 2016-03-18 01:59 output3/_SUCCESS
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:59 output3/part-v003-o000-r-00000
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:59 output3/part-v003-o000-r-00001
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:59 output3/part-v003-o000-r-00002
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:59 output3/part-v003-o000-r-00003
-rw-r--r-- 3 root hdfs 3 2016-03-18 01:59 output3/part-v003-o000-r-00004 which means with PARALLEL it creates multiple files within the same directory. In terms of MultiStorage, it created a separate directory and separate file. Additionally with MultiStorage you can pass compression, granted it's bz2, gz, no snappy and delimiter. It's clunky and documentation is not the best but if you need that type of control, it's an option.
... View more
03-11-2016
03:05 AM
1 Kudo
@Arun Sundar Selvan thanks for confirming that. We knew it worked with port 8032, just want to be clear you understand Oozie Spark action in HDP is not supported at this time. There are workarounds but nothing we put a stamp of approval on yet.
... View more
03-10-2016
10:46 PM
1 Kudo
Even if something did get deployed, you can reset and rerun the install. Once nodes register with ambari server, Ambari will report that it found binaries for hadoop and advise to run cleanup script. Additionally you could yum erase anything hdp.
... View more
03-10-2016
01:52 PM
@Sridhar Babu M is your hdfs up? Please make sure your cluster is up and passes service checks.
... View more