Member since
04-24-2016
52
Posts
11
Kudos Received
0
Solutions
06-15-2016
06:13 AM
1 Kudo
After built Atlas, I just only set the ATLAS_HOME_DIR in atlas-env.sh, and other settings, in atlas-env.sh and atlas-application.properties, are default. I try to import metadata according http://atlas.apache.org/Bridge-Hive.html After set $HIVE_CONF_DIR, I found that I can't set following configuration in atlas-application.properties. <property>
<name>atlas.cluster.name</name>
<value>primary</value>
</property> This is a XML style code, but the atlas-application.properties is not XML style, so I can't add this. I am wondering if the official guide of atlas is not accurate? Then I skip this setting, run import-hive.sh. It showed following: Exception in thread "main" org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException:
Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
Caused by: java.lang.reflect.InvocationTargetException
Caused by: javax.jdo.JDOFatalInternalException: Error creating transactional connection factory
NestedThrowables:
java.lang.reflect.InvocationTargetException
Caused by: org.datanucleus.exceptions.NucleusException: Attempt to invoke the "BONECP" plugin to create a ConnectionPool gave an error : The specified datastore driver ("com.mysql.jdbc.Driver") was not found in the CLASSPATH. Please check your CLASSPATH specification, and the name of the driver.
Caused by: org.datanucleus.store.rdbms.connectionpool.DatastoreDriverNotFoundException: The specified datastore driver ("com.mysql.jdbc.Driver") was not found in the CLASSPATH. Please check your CLASSPATH specification, and the name of the driver.
in order to import metadata into Atlas, what could I do next ?
... View more
Labels:
06-15-2016
05:22 AM
It works. Thank you very much.
... View more
06-14-2016
01:10 PM
Thank you very much. It works. But when I visited the default homepage, it required login. How can I set other username and password of login?
... View more
06-13-2016
11:04 AM
According these guide http://atlas.apache.org/InstallationSteps.html and http://atlas.apache.org/Configuration.html, after depressing Atlas 0.7.tar.gz, I just only set ATLAS_HOME_DIR and other settings are default. I ran "bin/atlas_start.py" to start atlas server. Then I ran "bin/quick_start.py" and it showed "Example data added to Apache Atlas Server!" successfully. When I visited http://localhost:21000/api/atlas/admin/version, it showed a login interface. Then I type "admin" as username and password. it showed the version of atlas successfully. But when I try to visit http://localhost:21000/dashboard, it shows 404 ERROR. I read the official guide or documentation many times, but I can't visit the dashboard in browser. Please help me. Thank you very much.
... View more
Labels:
05-06-2016
12:43 AM
Thank you very much.
... View more
05-06-2016
12:38 AM
Thank you very much. Could you leave your email address? So I can send email to you if I have some questions in the future.
... View more
05-03-2016
07:29 AM
I am learning Apache Atlas and read the demo on https://github.com/shivajid/atlas/blob/master/tutorial/Step1.md. This tutorial shows that import metadata into Atlas without using Kafka. So, Kafka is not necessary when we want to import metadata from Hive to Atlas, right? And from the docs of http://atlas.apache.org/Architecture.html, I realize that there will have some failure of communication between Hook and Atlas because of network issues and hence inconsistent of metadata. The Kafka would avoid this inconsistency happen, right? What other function or benefit if we use Kafka in the communication between Hook and Atlas? Thank you very much.
... View more
Labels:
- Labels:
-
Apache Atlas
-
Apache Hive
-
Apache Kafka
05-03-2016
07:11 AM
I can not understand the difference between Bridge and Hook completely. I think they are similar and they both are the middleware between Atlas and Other components. And I also watch a video on youtube which shows that they are not completely same. From this screenshot, the Hive Bridge seems relative with Client, and the Hive Hook seems relative with Post execution? Could anyone tell me what's the difference between Bridge and Hook specifically? And What situation I should use Bridge? What situation I should use Hook? Thank you very much.
... View more
Labels:
05-03-2016
06:43 AM
Do you mean that I must shutdown Atlas server before running the quick_start.py script? And there is another important question: where does quick_start.py script get data from? In the other words, where should I import my data before running this script?
... View more
05-02-2016
01:44 AM
I am a student and these questions make me frustrated several weeks, please help me. According a developer in this community, running Atlas needs Hive, namely, the minimum system of Atlas must consist of Atlas, Hive, Hadoop and a RDBMS such as MySQL. I also did much research, and read the tutorial on https://github.com/shivajid/atlas. But this tutorial is based on HDP, and my teacher requires me to configure the minimum system by myself. I imitated this tutorial of shivajid/atlas mentioned above, and summary some main steps of this tutorial: Step1: Download Hadoop and configure it. Step2: Execute git clone to get the source code of Atlas, build it by maven, and configure it. ### After this step, I started Atlas server, in order to verify if the server is up and running, I used a web browser to visit " ### http://localhost:21000/api/atlas/admin/version", and it showed some version information. but when I tried to visit " ### http://localhost:21000" or "http://localhost:21000/dashboard", it showed HTTP ERROR 500. This is my first question. Step3: Install MySQL-server, create MySQL user and grant privileges to him. Step4: Get binary Hive and configure MySQL as the metadata repository of Hive. Step5: Execute the script of "MySQLSourceSystem.dll" which comes from tutorial of shivajid/atlas on github. Step6: Import data from "drivers.csv" and "timesheet.csv" to tables in MySQL. Step7: Execute script quick_start.py which in ${atlas_dir}/bin. ### But it report that, no data was added to the Apache Atlas Server. This is my second question and I stop here now. The following steps,which were explained in the tutorialI, haven't been done: Step8: Execute script sqoop_job.sh Step9: Validate tables in Hive Step10: Check metadata in Atlas Web UI. Summary: I did much research to configure the minimun system of Atlas but failed, so I concluded my main steps shown above. But these are still a lots of problems: 1. After started Atlas server (Step 2), why I had failed to visit "http://localhost:21000/dashboard" and got ERROR 500. 2. After execute quick_start.py (Step 7), why it report that no data was added to the Apache Atlas Server? 3. I think, the tutorial of shivajid/atlas on github is outdated which was post in Sep 2015, and it must use HDP sanbox. Could anyone give me or help me to write a brief tutorial of minimum system of Atlas which just conclude these component: Atlas, Hive, Hadoop, MySQL. This is my Email: dreamcoding@hotmail.com, If anyone have some relative stuff or material, please send them to me. Thank you very much. Thank you!
... View more
Labels:
- « Previous
- Next »