Member since
07-10-2017
78
Posts
6
Kudos Received
4
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
4435 | 10-17-2017 12:17 PM | |
7004 | 09-13-2017 12:36 PM | |
5390 | 07-14-2017 09:57 AM | |
3514 | 07-13-2017 12:52 PM |
07-17-2017
05:25 AM
Hello @Saurab Dahal , You should use the hive context object to connect and query hive Here in the answer you can see a small example https://community.hortonworks.com/questions/93392/how-to-connect-and-run-hive-query-from-apache-spar.html Michel
... View more
07-17-2017
05:00 AM
Hi, I do t know for the snapshot but you can have the same result by doing an export and the. And import of your table. In the following link you have an example how to do it: http://www.rotanovs.com/hbase/hbase-importexport/ Michel
... View more
07-14-2017
04:23 PM
You can add compression when you write your data. This will speed up the saving because the size of the data will smaller. Also increase the number of partition
... View more
07-14-2017
02:09 PM
Hi, I started the metron-ui and rest-api from commande line as explained in the HCP doc. What's the clean way (command) to stop the metron-ui and the rest-api? Thanks, Michel
... View more
Labels:
- Labels:
-
Apache Metron
-
Apache Ranger
07-14-2017
01:28 PM
Hi,
You just need put a processor "puthdfs", then righ click to open the configuration panel. they you need to provide the configuration file of hdfs (hdfis-site.xml, core-site.xml) and the folder where you want to put the data. This is to put data to hdfs. To read data (for example locally) use the processor tailfile.
... View more
07-14-2017
09:57 AM
Hi,
You can do it, by create a simple connection to hdfs with hdfs client.
For example in Java, you can do the following: Configuration confFS = new Configuration();
confFS.addResource("/etc/hadoop/conf/core-site.xml");
confFS.addResource("/etc/hadoop/conf/hdfs-site.xml");
FileSystem dfs2 = FileSystem.newInstance(confFS);
Path pt = new Path("/your/file/to/read");
BufferedReader br = new BufferedReader(new InputStreamReader(dfs2.open(pt)));
String myLine;
while ((myLine = br.readLine()) != null) {
System.out.println(myLine);
}
br.close();
dfs2.close();
This code will create a single connection to hdfs and read a file defined in the variable pt
... View more
07-14-2017
09:39 AM
Hi @Adnan Alvee,
In order to parrallelize the write to hdfs, you just need to increase the number of partitions for your data and or increase the number of executor. To increase the number of executor, when you submit your spark job specify the option: --num-executors x where the x is the number of executor that you want. more you have more parallelism your have. to increase the number of partition, in your code, you have to call the function repartition(x) on the RDD or the Dataset, it will spread the data over x node (container) and each node will write in parallel. Last thing, don't increase to much the repartition because the result can be that you create too many small file. So I advice, Size of full data (in MB) / 128 = repartition number Michel
... View more
07-14-2017
08:50 AM
Hi @asubramanian, Thanks for your reply. I followed the HCP documentation and it's never mention that the user metron should run the process. Moreover in the documentation it's specify to install and run the rest-api and metron-UI manually in command line. HCP doc Do you have any docuementation or tutorial to run rest-api and metron-ui with ambari? In my case, If I run the rest-api and metron-ui with metron user I will have the same problem, which is a permission denied when I try to create a new sensor, because the hdfs path is /apps/metron/patterns has 755 right access. Should I change the owner of the folder to metron? Thanks, Michel
... View more
07-13-2017
09:28 PM
same question as that one:
https://community.hortonworks.com/questions/113885/hive-table-on-hbase-table-showing-null-for-integer.html
... View more
07-13-2017
03:29 PM
Hi,
Normally which user should start the process rest-api and the metron UI? Is it root, metron or hdfs?
I'm asking that because if it's root, then we are not able to create a new sensor because root doesn't have the right access to write in the folder /apps/metron/patterns (in hdfs). Only the user hdfs have the right to write their. A work around will be to do a chmod on the parent folder, but it's not really secure. So the question is who should run/start the process metrion UI and rest api? Thanks,
Michel
... View more
Labels:
- Labels:
-
Apache Metron
- « Previous
- Next »