Member since
03-07-2019
158
Posts
53
Kudos Received
33
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
6359 | 03-08-2019 08:46 AM | |
4341 | 10-17-2018 10:25 AM | |
2774 | 10-16-2018 07:46 AM | |
2121 | 10-16-2018 06:57 AM | |
1763 | 10-12-2018 09:55 AM |
07-25-2018
02:49 PM
I believe FetchParquet does what you need; https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-parquet-nar/1.5.0/org.apache.nifi.processors.parquet.FetchParquet/index.html
... View more
07-24-2018
04:09 PM
@Gayathri Devi I've created a database in MariaDB and exported a hive table using sqoop on my lab setup. This worked well for me; [sqoop@jsneep-lab ~]$ sqoop export --connect jdbc:mysql://172.3.2.1/export --username mariadb --password mariadb --table exported --direct --export-dir /apps/hive/warehouse/drivers Make sure you have /usr/share/java/mysql-connector-java.jar present on your system, this gave me trouble initially.
... View more
07-24-2018
08:23 AM
Hi @Anurag Mishra Prior to HDP 3 you could only see that an application was killed by a user, not who killed the application HDP 3 and onwards is more informative about who killed an application. [jsneep@node4 ~]$ yarn jar /usr/hdp/2.6.4.0-91/hadoop-mapreduce/hadoop-mapreduce-examples.jar pi 10 9000
18/07/24 07:44:44 INFO security.TokenCache: Got dt for hdfs://hwc1251-node2.hogwarts-labs.com:8020; Kind: HDFS_DELEGATION_TOKEN, Service: 172.25.33.145:8020, Ident: (HDFS_DELEGATION_TOKEN token 7 for jsneep)
18/07/24 07:44:45 INFO input.FileInputFormat: Total input paths to process : 10
18/07/24 07:44:45 INFO mapreduce.JobSubmitter: number of splits:10
18/07/24 07:44:45 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1532417644227_0005
18/07/24 07:44:46 INFO impl.YarnClientImpl: Submitted application application_1532417644227_0005
18/07/24 07:44:46 INFO mapreduce.Job: Running job: job_1532417644227_0005
[root@hwc1251-node4 ~]# yarn application -kill application_1532417644227_0005
18/07/24 07:44:53 INFO mapreduce.Job: Job job_1532417644227_0005 failed with state KILLED due to: Application killed by user.
18/07/24 07:44:53 INFO mapreduce.Job: Counters: 0
Job Finished in 8.516 seconds
Ex, above I've submitted a yarn job (application_1532417644227_0005) & killed it. The logs state "Application killed by user." I can also browse the Resource Manager UI at http://<RM IP ADDRESS>:8088/cluster/apps/KILLED and see that it was killed by a user. The apache jira for this: https://issues.apache.org/jira/browse/YARN-5053 | "More informative diagnostics when applications killed by a user" In my HDP3 cluster, when I submit an identical job and kill it; [root@c2175-node4 ~]# yarn app -kill application_1532419910561_0001
18/07/24 08:12:45 INFO client.RMProxy: Connecting to ResourceManager at c2175-node2.hwx.com/172.25.39.144:8050
18/07/24 08:12:45 INFO client.AHSProxy: Connecting to Application History server at c2175-node2.hwx.com/172.25.39.144:10200
Killing application application_1532419910561_0001
For example now via the RM UI, I can browse to http://<RM IP ADDRESS>:8088/ui2/#/yarn-app/application_x_/info and under diagnostics we will see the user and source address of the kill operation. The same would be visible through CLI, via "yarn app -status application_1532419910561_0001 | grep killed" Application application_1532419910561_0001 was killed by user root at 172.25.33.15 Edit: PS, you could make use of YARN Queues & ACLs to limit / determine who has rights to kill yarn applications, I wanted to mention this in case you're looking for something to help you if you're currently unable to get your cluster upgraded to HDP3. Further info: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.5/bk_yarn-resource-management/content/controlling_access_to_queues_with_acls.html *please mark this answer as accepted if you found this helpful*
... View more
07-23-2018
04:19 PM
@Sambasivam Subramanian I've just noticed that there is a script to help on this available, which I was unaware of, it would be easier this way; [root@lab-1 ~]# /usr/hdp/2.6.5.0-292/zeppelin/bin/install-interpreter.sh -n python
Install python(org.apache.zeppelin:zeppelin-python:0.7.0) to /usr/hdp/2.6.5.0-292/zeppelin/interpreter/python ...
Interpreter python installed under /usr/hdp/2.6.5.0-292/zeppelin/interpreter/python.
1. Restart Zeppelin
2. Create interpreter setting in 'Interpreter' menu on Zeppelin GUI
3. Then you can bind the interpreter on your note
... View more
07-23-2018
02:08 PM
Hi @Sambasivam Subramanian Yes, you definitely need to have the python interpreter data under /usr/hdp/current/zeppelin-server/interpreter, you can download zeppelin here and extract the python interpreter from the apache download. Also be sure to update interpreter.json afterwards.
... View more
07-23-2018
01:54 PM
@Gayathri Devi Perhaps this works, i haven't tried this myself yet as I don't currently have any lab setup using mariaDB. Edit; This may also be helpful: https://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_syntax_3 I've noticed now the --export-dir is also required; sqoop export \
--connect "jdbc:mariadb://localhost/example" \
--username mariadb \
--password mariadb \
--table hivetable \ --export-dir /apps/hive/warehouse/hivetable
... View more
07-23-2018
01:04 PM
1 Kudo
Hi @Gayathri Devi it seems you're trying to use -hcatalog-table
The argument value for this option is the HCatalog tablename. The presence of the --hcatalog-table option signifies that the import or export job is done using HCatalog tables, and it is a required option for HCatalog jobs.
But it looks like you have an accidental space between the -hcatalog and -table, resulting in a command syntax issue; https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.5/bk_data-movement-and-integration/content/sqoop_hcatalog_integration.html
... View more
07-20-2018
03:33 PM
Hi @Sambasivam Subramanian Can you share all the steps you've taken? Note that zeppelin on HDP supports these, and that using pyspark you can run python scripts;
Spark JDBC (supports Hive, Phoenix) OS Shell Markdown Livy (supports Spark, Spark SQL, PySpark, PySpark3, and SparkR) AngularJS source: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.5/bk_zeppelin-component-guide/content/using-interp.html [root@c1175-node3 zeppelin-server]# cd interpreter/
[root@c1175-node3 interpreter]# ll
total 44
drwxr-xr-x 3 zeppelin zeppelin 4096 Jul 20 15:10 angular
drwxr-xr-x 3 zeppelin zeppelin 20480 Jul 20 15:10 jdbc
drwxr-xr-x 3 zeppelin zeppelin 4096 Jul 20 15:10 lib
drwxr-xr-x 3 zeppelin zeppelin 4096 Jul 20 15:10 livy
drwxr-xr-x 3 zeppelin zeppelin 4096 Jul 20 15:10 md
drwxr-xr-x 3 zeppelin zeppelin 4096 Jul 20 15:10 sh
drwxr-xr-x 7 zeppelin zeppelin 4096 Jul 20 15:11 spark
It is however also possible to add to this manually, did you download zeppelin manually and copy the interpreter/python directory to your HDP install of zeppelin, as part of the steps you've undertaken? Please also note that you need to update interpreter.json which is normally found under /usr/hdp/current/zeppelin-server/conf
... View more
07-20-2018
09:56 AM
There are indeed 2 addresses, the sandbox is Centos which has an IP, and in centos there is a docker image running HDP which has another IP. From windows itself, ports on the docker image are forwarded and accessible via localhost. Ex when your sandox is up, you could access 127.0.0.1:8888 via a browser. Similarly port 2222 should also be forwarded to be accessible through localhost/127.0.0.1.
... View more
07-20-2018
07:39 AM
@Tulasi Uppu Are you using the IP address that you've seen via docker inspect? Try the ip visible via docker-machine ls And also be sure to check that windows firewall isn't preventing your access 🙂 Let me know if that solves your issue or if you have more questions! If this has helped, please take a moment to login and mark this answer as accepted!
... View more
- « Previous
- Next »