Member since
09-10-2015
95
Posts
166
Kudos Received
34
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1358 | 11-04-2016 04:56 PM | |
1118 | 10-21-2016 07:13 PM | |
2233 | 03-09-2016 07:00 PM | |
2595 | 01-28-2016 12:27 AM | |
1420 | 12-10-2015 03:09 PM |
04-05-2021
10:51 PM
Hi, You must deploy the jce policy on every cluster nodes if you are using built-in openjdk. Please follow the steps. [root@hostname]# locate local_policy.jar [root@hostname]# wget --no-check-certificate --no-cookies --header "Cookie: oraclelicense=accept-securebackup-cookie" "http://download.oracle.com/otn-pub/java/jce/8/jce_policy-8.zip" [root@hostname]# unzip -o -j -q jce_policy-8.zip -d /usr/jdk64/jdk1.8.0_112/jre/lib/security/ [root@hostname]# ambari-server rstart Now test the connection. Hopes it will resolves the issue.
... View more
01-27-2021
12:00 AM
Hi @abajwa, Does the Ambari Server Host also need to present it's own SSL certificate to the AD server? In case of multiple domain controllers, do we need to have separate SSL certificates from each of the domain controllers? Thanks, Megh
... View more
03-23-2020
02:14 AM
hello, This wokarround didn't work for me. Configured the LDAP setup so that the BaseDN matches only 1 entry. calling "ambari-server sync-ldap --existing" didn't remove all existing LDAP Users and groups rather it deleted 2 only. may be i missed something, but after running the setup do we need to restart ambari-server? What should be the expected behaviour when runing the "ambari-server sync-ldap --all" and the BaseDN pointing to a single AD entry? The doc states the following for option '--exisiting' : "Users will be removed from Ambari if they no longer exist in LDAP, and group membership in Ambari will be updated to match LDAP". Since AD users still exist that would have no effect to remove the users even if baseDN points to single entry. What we are looking for (HDP2.6.5) is to remove all LDAP synced users other than these specified in --users users.txt and --groups group.txt. It looks like there is no such tool and we have to resort to manually use ambari APIs somehow. One thing i'm not sure is how are the lowercased alias being handled, since during the first sync we had the default value 'true' to force lower case, and now changed it to 'false' looking forward your insights
... View more
01-23-2020
10:31 AM
Thanks it really fixes the problem RHEL 6.10 (Open JDK 1.8)
... View more
07-13-2018
06:59 PM
5 Kudos
A common question when planning a new implementation or an upgrade is “what’s supported?” We’ve traditionally handled this with documentation, but as we release more products and support more platforms, it’s been tough for our customers to understand product interoperability. To solve this problem, we’ve created a new interactive web application that Hortonworks Support users can use to quickly answer questions like the following: Which version of Ambari works with HDF 3.1.2? Which Hortonworks products are certified with RHEL 7.4? Which databases are supported with both HDP 2.6.4 and HDF 3.1.2? Usually, these questions would be tough to answer as multiple documentation sections would have to be referenced and compared, but with the Hortonworks Support Matrix, answers are only a few clicks away. Just navigate to https://supportmatrix.hortonworks.com and log in using your Hortonworks Support Portal credentials; then click one or more products to see what we support. It’s important to note that you must have a Hortonworks Support Portal account to access this information.
... View more
07-15-2019
01:20 PM
Hi everyone! I installed the ambari server and ambari-agent to my laptop (SuSE Tumbleweed 2019) to set up a local single-node Hadoop cluster. (Server and agent versions 2.1.2.1-418) Server starts without issues, and I can open the console at http://localhost:8080 Agent says it started, but it can't connect the server, here's what the log says : ERROR 2019-07-15 11:06:06,862 NetUtil.py:77 - [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:727)ERROR 2019-07-15 11:06:06,862 NetUtil.py:78 - SSLError: Failed to connect. Please check openssl library versions.Refer to: https://bugzilla.redhat.com/show_bug.cgi?id=1022468 for more details. I have python (2.7.16-2.1) installed. Also python3 is installed, but the executable defaults to version 2.7 : ls -l /usr/bin/python lrwxrwxrwx 1 root root 9 Jun 24 22:12 /usr/bin/python -> python2.7 I would implement this "verify=disable" solution mentioned above, but there is no file named cert-verification.cfg in my system. There is no python dir under etc at all ! So I think maybe python under SuSE might have a different configuration than other distributions. I google'd a lot, but couldn't find a solution for this yet. Can someone guide me thru this issue please ?
... View more
07-06-2016
04:24 PM
13 Kudos
Hortonworks Data Platform Artifacts Developing solutions with Hadoop commonly requires the use of multiple different HDP component libraries. Whether you’re building solutions with Pig, Spark, Cascading, or HBase, at some point extensions will need to be created, and those artifacts for each component will need to be used.
This guide serves as an overview of where to find those artifacts and how to get them quickly integrated with your preferred build tool, and IDE. Reposotories At Hortonworks, we store all of our artifacts in a public Sonatype Nexus repository. That repository can be easily accessed and searched for commonly used library, source code, and javadoc archives simply by navigating to http://repo.hortonworks.com. Artifacts Jar files containing compiled classes, source, and javadocs are all available in our public repository, and finding the right artifact with right version is as easy as searching the repository for classes you need to resolve. For example, If creating a solution that requires the use of a class such as org.apache.hadoop.fs.FileSystem, you can simply search our public repository for the artifact that contains that class using the search capabilities available through http://repo.hortonworks.com. Searching for that class will locate the hadoop-common artifact that is part of the org.apache.hadoop group. There will be multiple artifacts each with a different version. Artifacts in our repository use a 7 digit version scheme. So if we’re looking at the 2.7.1.2.3.2.0-2650 version of this artifact:
The first three digits (2.7.1) signify the Apache Hadoop base version The next four digits (2.3.2.0) signify our Hortonworks Data Platform release The final numbers after the hyphen (2650) signifies the build number As you’re looking for the right artifact, it’s important to use the artifact version that corresponds to the HDP version you plan to deploy to. You can determine this by using hdp-select versions from the command line, or using Ambari by navigating to Admin > Stack and Versions. If neither of these are available in your version of HDP or Ambari, you can use yum, zypper, or dpkg to query the RPM or Debian packages installed for HDP and note their versions. Once the right artifact has been found with the version that corresponds to your target HDP environment, it’s time to configure your build tool to both resolve our repository and include the artifact as a dependency. The following section outlines how to do both with commonly used with build tools such as Maven, SBT, and Gradle. Maven Setup Apache Maven, is an incredibly flexible build tool used by many Hadoop ecosystem projects. In this section we will outline what updates to your project’s pom.xml file are required to start resolving HDP artifacts. Repository Configuration The pom.xml file enables flexible definition of project dependencies and build procedures. To add the Hortonworks repository to your project, allowing HDP artifacts to be resolved, edit the <repositories/> section and add a <repository/> entry as illustrated below: <repositories>
<repository>
<id>HDP</id>
<name>HDP Releases</name>
<url>http://repo.hortonworks.com/content/repositories/releases/</url>
</repository>
</repositories>
Artifact Configuration Dependencies are added to Maven using the <dependency/> tag within the <dependencies/> section of the pom.xml. To add a dependency such as hadoop-common, add this fragment: <dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-common</artifactId>
<version>2.7.1.2.3.2.0-2650</version>
</dependency>
Once both the repository has been added to the <repositories/> section, and the artifacts have been added to the <dependencies/>, a simple mvn compile can be issued from the base directory of your project to ensure that proper syntax has been used and the appropriate dependencies are downloaded. Source & Javadoc When using Maven with an IDE, it is often helpful to have the accompanying JavaDoc and source code. To obtain both from our repository for the artifacts that you have defined in your pom.xml, run the following commands from the base directory of your project: mvn dependency:sources
mvn dependency:resolve -Dclassifier=javadoc SBT Setup The Scala Build Tool is commonly used with Scala based projects, and provide simple configuration, and many flexible options for dependency and build management. Repository Configuration In order for SBT projects to resolve Hortonworks Data Platform dependencies, an additional resolvers entry must be added to your build.sbt file, or equivalent, as illustrated below: resolvers += "Hortonworks Releases" at "http://repo.hortonworks.com/content/repositories/releases/" Artifact Configuration Dependencies can be added to SBT’s libraryDependencies as illustrated below: libraryDependencies += "org.apache.hadoop" % "hadoop-common" % "2.7.1.2.3.2.0-2650" To explicitly ask SBT to also download source code and JavaDocs an alternate notation can be used: libraryDependencies += "org.apache.hadoop" % "hadoop-common" % "2.7.1.2.3.2.0-2650" withSources() withJavadoc() Once both the repository has been added to resolvers, and the artifacts have been added to dependencies, a simple sbt compile can be issued from the base directory of your project to ensure that proper syntax has been used and the appropriate dependencies are downloaded. Gradle Setup The Gradle build management tool is used frequently in Open Source java projects, and provides a simple Groovy-based DSL for project dependency and build definition. Plugin Configuration Gradle uses plugins to add functionality to add new task, domain objects and conventions to your gradle build. Add the following plugins to your build.gradle file, or equivalent, as illustrated below: apply plugin: 'java'
apply plugin: 'maven'
apply plugin: 'idea' // Pick IDE appropriate for you
apply plugin: 'eclipse' // Pick IDE appropriate for you
Repository Configuration In order for Gradle projects to resolve Hortonworks Data Platform dependencies, an additional entry must be added to your build.gradle file, or equivalent, as illustrated below: repositories {
maven { url "http://repo.hortonworks.com/content/repositories/releases/" }
}
Artifact Configuration Dependencies can be added to Gradle’s dependencies section as illustrated below: dependencies {
compile group: "org.apache.hadoop", name: "hadoop-common", version: "2.7.1.2.3.2.0-2650"
}
idea { // Pick IDE appropriate for you
module {
downloadJavadoc = true
downloadSources = true
}
}
eclipse { // Pick IDE appropriate for you
classpath {
downloadSources = true
downloadJavadoc = true
}
}
Once both the repositories and the dependencies have been added to build file, a simple gradle clean build can be issued from the base directory of your project to ensure that proper syntax has been used and the appropriate dependencies are downloaded.
... View more
08-19-2016
10:25 PM
where and how can I add this parameter? I am having same issues. thanks
... View more
11-09-2015
09:24 PM
Which kerberos wizard option did you use MIT, AD, or Manual?
... View more
04-21-2017
10:28 AM
The unexpected benefit of this is that nobody will ever forget the LDAP password again: not will it be included in your favourite shell's history file, but anyone who can log in on that node will also be able to see those options by keeping an eye on ps. Isn't that neat? Don't do this, kids. Never write passwords on the command line.
... View more