Steps for importing the oracle DB in Atlas to show lineage.
To run an Oracle import you must have the Oracle driver in the sqoop path. Make sure you also have an entry in the /etc/hosts for the remote Oracle server and the IP should be reachable or hostname resolution should work.
The user executing the import should have the privileges to the undelaying oracle table, you can Login to the Oracle instance on the Sqoop command line as below:
--connect jdbc:oracle:thin:@OracleServer:OraclePort:OracleInstance --username UserName -P
sqoop will process the below command command in its console output, and then run the MapReduce jobs to bring in the data via the Oracle JDBC driver, the imported data will appaer as a series of “part-m” files, CSV files with one per MapReduce reducer.
$ sqoop import --connect jdbc:oracle:thin:@localhost:1521/orcl --username oracleapps --password oracle1234 --table customer
By default, sqoop will put the resulting file in your user’s home directory in HDFS e.g /user/sqoop You can also import Oracle and other RDBMS data directly into Hive, with sqoop creating equivalent datatypes for the data coming in (basic datatypes only, none of the advanced spatial and other Oracle ones). For example, I could import the customer table in the orapps schema in like this, directly into an equivalent Hive table:
$ sqoop import --connect jdbc:oracle:thin:@localhost:1521/orcl --username orapps --password oracle1234 --table customer --hive-import<
Execute from hive CLI
$ hive hive> desc customer; $ sqoop import --connect jdbc:oracle:thin:@localhost:1521/orcl --username orapps --password oracle1234 --table customer --hive-import --incremental append --check-column Cust_ID