By definition, Atlas provides a
scalable and core foundational services for Data Governance - enabling
enterprises to efficiently and effectively meet their compliance requirements
with their Hadoop eco system. However, it is a complex application, which is
built using the integration of various components in the Hadoop eco system.
Below are the components involved in the Architecture:
The intention of this article is to provide
troubleshooting tips if Ambari install is not functioning correctly.
Install validation and troubleshooting tips:
Make sure Atlas Metadata Server and Atlas
Metadata clients are installed from Ambari.
Install Ranger Tagsync component if you wish to
do authorization control using Atlas Tags in Ranger.
Make sure the below Kafka topics are created in
You can check this by
using the following command on any of the kafka brokers: (In Kerberized cluster
you need kafka key tab to run this command)
If the Kafka topics
didn’t get created, you can create them manually using the attached
atlas_kafka_acl.sh script. All you need to do is to update the Zookeeper quorum
in the script and run it on any Kafka broker.
Depending upon whether
you have Ranger in your environment and whether Kafka topic authorization is
controlled by Ranger, You should see necessary policies created in Ranger Kafka
repository in Ranger. If you don’t find those policies, you need to create
policies in Ranger Kafka repository in Ranger granting necessary accesses to
ATLAS_ENTITIES and ATLAS_HOOK topics. Below are how the policies should be set
up: in Ranger —> Resource based policies —> <Cluster Name>_kafka