Member since
05-30-2018
1322
Posts
715
Kudos Received
148
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 4051 | 08-20-2018 08:26 PM | |
| 1952 | 08-15-2018 01:59 PM | |
| 2377 | 08-13-2018 02:20 PM | |
| 4111 | 07-23-2018 04:37 PM | |
| 5021 | 07-19-2018 12:52 PM |
08-11-2016
03:06 PM
I am planning to non default hdp service accounts, what should I know? for example the default service account for HDFS use is hdfs. I want to change to hdfsuser. I may use non defaults for all service account. Any gotcha I should be aware of?
... View more
Labels:
08-11-2016
02:40 PM
2 Kudos
@mike harding check if you have tez reuse container turned on tez.am.container.reuse.enabled=true Configuration that specifies whether a container should be reused. This allow other application to reuse tez containers to increase performance. turn it off if you are not interested in that functinoality.
... View more
08-11-2016
01:52 PM
2 Kudos
If i understand ur question and i am making a assumption here, you want to store the pdf in hive to perform some analysis on the content inside the pdf file. The way i have done this is with apache Tika. The process will extract text file pdf and store in text format file. Then apply data model to text and store in hive.
... View more
08-10-2016
07:13 PM
1 Kudo
@SBandaru Per documentation: Alternately, you can use the pre-2.2 method:
Link the HBase configuration file with the Phoenix libraries: ln -sf <HBASE_CONFIG_DIR>/hbase-site.xml /usr/hdp/current/phoenix-client/bin/hbase-site.xml Link the Hadoop configuration file with the Phoenix libraries: ln -sf <HADOOP_CONFIG_DIR>/core-site.xml /usr/hdp/current/phoenix-client/bin/core-site.xml ln -sf >HADOOP_CONFIG_DIR>/hdfs-site.xml /usr/hdp/current/phoenix-client/bin/hdfs-site.xml
... View more
08-10-2016
03:07 PM
1 Kudo
If imported to Hive directly, will the table be available to Pig, MR ? Yes since hive and pig are integrated with HS2 If imported to HCatalog, what needs to be done for accessing via Hive ? Nothing. you should see the table in hive Is pre-creation of tables in Hive necessary ? If yes, what is the advantage of the importing in HCatalog, (I can import in Hive directly)/(import in HDFS and then create external table) ? No. you can use --create-hive-table arguments
... View more
08-09-2016
02:54 PM
1 Kudo
@Khera for all intents and purposes stop using hive CLI. It will soon be deprecated in favor of beeline. Jira here. It does not integrate with ranger so it bypass the security policies. I recommend you only use beeline.
... View more
08-09-2016
02:48 PM
2 Kudos
@Johnny Fugers Hive is great for typically BI queries. The scalability is limitless. When you get into the area of updates, I rather do those activities on phoenix and serve the end results back to hive for BI queries. Hive ACID is coming soon. Until that is available I would use the phoenix->Hive route. Use PIG for ETL. Where it gets interested is using a MPP database on Hadoop. that is where HAWQ comes in. It is a good low latency db engine which provided you some benefits from both hive and phoenix. It does not do all hive & phoenix capabilities. I would say it is a good happy medium. I hope that helps. When you go further into your journey you will start to ask question about security and governance. For security you will start with ranger & Knox. and goverance you will start with falcon/atlas/ranger.
... View more
08-09-2016
04:29 AM
@subacini balakrishnan did you set delete.topic.enable=true for all brokers? some topic may be undeletable in .9. fixed in .9.1. jira here when you made the delete.topic.enable=true, did you restart ZK and Kafka brokers?
... View more
08-08-2016
04:11 PM
@Eric Brosch Phoenix is a SQL skin on top of hbase. Phoenix allows to create secondary index on hbase which hbase natively does not create. Phoenix on HDP comes out of the box with hbase.
... View more
08-08-2016
02:33 PM
1 Kudo
@Eric Brosch Phoenix and HDB (HAWQ) may be leveraged in a Lambda architecture. Phoenix supports secondary index and HAWQ is a relational MPP db on HDP. Both can serve low latency queries. Choosing between the two? For know query patterns Phoenix will perform well. For unknown query patterns HAWQ may be your way to go
... View more