Member since
03-14-2016
4721
Posts
1111
Kudos Received
874
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2534 | 04-27-2020 03:48 AM | |
| 5003 | 04-26-2020 06:18 PM | |
| 4095 | 04-26-2020 06:05 PM | |
| 3301 | 04-13-2020 08:53 PM | |
| 5042 | 03-31-2020 02:10 AM |
03-29-2017
05:51 AM
@Sanjib Behera It will be good if you can mark this HCC thread as Answerd by clicking on the "Accept" button on the correct answer that helps others.
... View more
03-29-2017
05:40 AM
@Sanjib Behera
Good to know that previously mentioned suggestion worked. Regarding the " spoutConfig.securityProtoco " config you can refer to the following link to know more about it. Sometimes an incorrect value to this property causes issue, but in your case looks like it is alredy running fine. https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.2/bk_storm-user-guide/content/stormkafka-secure-config.html
... View more
03-29-2017
05:28 AM
@heta desai
The following doc will give mode detail on this: 1. https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.5.3/bk_command-line-installation/content/determine-hdp-memory-config.html 2. https://hortonworks.com/apache/mapreduce/#section_1 3. https://hortonworks.com/blog/how-to-plan-and-configure-yarn-in-hdp-2-0/ You can specify the minimum unit of RAM to allocate for a Container. The tasks are run within containers launched by YARN. mapreduce.{map|reduce}.memory.mb is used by YARN to set the memory size of the container being used to run the map or reduce task. If the task grows beyond this limit, YARN will kill the container. .
... View more
03-29-2017
03:41 AM
@Jonathan T
Instead of shutting down the VM completely i usually do "Save State" because there are many components running inside the VM and it takes some time to get all those services restarted. But it depends on the usecase. If we have some time then sometimes it is better to restart the VM that way it is comparatively bit faster.
... View more
03-29-2017
02:56 AM
@Qinglin Xia You might be interested in the Nifi GetHDFSEvents processor. https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi.processors.hadoop.inotify.GetHDFSEvents/ . I personally did not check the UNLINK event though but i see a good example to validate the same. I try it you also might want to check the same in parallel: https://github.com/onefoursix/hdfs-inotify-example/blob/master/src/main/java/com/onefoursix/HdfsINotifyExample.java .
... View more
03-29-2017
02:11 AM
@Vishal Gupta You might want to refer to the detailed step by step guide on "Create Encryption Zone" In Lab6
https://github.com/HortonworksUniversity/Security_Labs#lab-6a
... View more
03-28-2017
07:02 PM
@n c
Hadoop DistCp (distributed copy) can be used to copy data between Hadoop clusters (and also within a Hadoop cluster). DistCp uses MapReduce to implement its distribution, error handling, and reporting. It expands a list of files and directories into map tasks, each of which copies a partition of the files specified in the source list. - You can use Falcon to continuously sync Hive tables between clusters. It uses "distcp" under the cover for the data but keeps track of Partitions added etc. Including secure ones being in the same realm makes it easier. Some setup is needed in the configuration to add both clusters to the environment of the one running the "distcp". https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_Sys_Admin_Guides/content/ch_distcp.html Some information if your clusters are HA: http://henning.kropponline.de/2015/03/15/distcp-two-ha-cluster/ And some info on Falcon Hive https://falcon.apache.org/HiveIntegration.html
.
... View more
03-28-2017
05:48 PM
@Keshav Piplani Have you followed the similar steps mentioned: https://community.hortonworks.com/articles/52876/enable-https-for-yarn-and-mapreduce2.html Do you see any error in the logs of those components? (or they are running fine) Do you see the communication on the https port? # openssl s_client -connect ${COMPONENTS}:${SECURE_PORT} -showcerts < /dev/null . Have you restarted ambari server. Browser cache is cleared (or tested form other browser)
... View more
03-28-2017
05:40 PM
3 Kudos
@PJ Are you setting: set hive.execution.engine=spark; Hive's execution engine only supports MapReduce & Tez. Running with
Spark is not supported in HDP at this current moment in time. https://issues.apache.org/jira/browse/HIVE-7292
... View more
03-28-2017
05:27 PM
@zkfs
Also have you performed the Hive MySQl setup as mentioned in the previously shared doc. like user creation and giving enough permission to the user. Example: Where <HIVEUSER> is the Hive user name, <HIVEPASSWORD> is the Hive user password and <HIVEMETASTOREFQDN> is the Fully Qualified Domain Name of the Hive Metastore host. # mysql -u root -p
CREATE USER '<HIVEUSER>'@'localhost' IDENTIFIED BY '<HIVEPASSWORD>';
GRANT ALL PRIVILEGES ON *.* TO '<HIVEUSER>'@'localhost';
CREATE USER '<HIVEUSER>'@'%' IDENTIFIED BY '<HIVEPASSWORD>';
GRANT ALL PRIVILEGES ON *.* TO '<HIVEUSER>'@'%';
CREATE USER '<HIVEUSER>'@'<HIVEMETASTOREFQDN>' IDENTIFIED BY '<HIVEPASSWORD>';
GRANT ALL PRIVILEGES ON *.* TO '<HIVEUSER>'@'<HIVEMETASTOREFQDN>';
FLUSH PRIVILEGES;
... View more