Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3557 | 05-03-2017 05:13 PM | |
| 2933 | 05-02-2017 08:38 AM | |
| 3183 | 05-02-2017 08:13 AM | |
| 3145 | 04-10-2017 10:51 PM | |
| 1621 | 03-28-2017 02:27 AM |
03-30-2016
11:40 PM
I got it to work with the following in my repo I linked earlier hdfs dfs -put drivers/* /tmp/udfs
beeline
!connect jdbc:hive2://localhost:10000 “” ””
add jar hdfs://sandbox.hortonworks.com:8020/tmp/udfs/mongo-hadoop-hive-1.5.0-SNAPSHOT.jar;
add jar hdfs://sandbox.hortonworks.com:8020/tmp/udfs/mongo-hadoop-core-1.5.0-SNAPSHOT.jar;
add jar hdfs://sandbox.hortonworks.com:8020/tmp/udfs/mongodb-driver-3.0.4.jar;
DROP TABLE IF EXISTS bars;
CREATE EXTERNAL TABLE bars
(
objectid STRING,
Symbol STRING,
TS STRING,
Day INT,
Open DOUBLE,
High DOUBLE,
Low DOUBLE,
Close DOUBLE,
Volume INT
)
STORED BY 'com.mongodb.hadoop.hive.MongoStorageHandler'
WITH SERDEPROPERTIES('mongo.columns.mapping'='{"objectid":"_id",
"Symbol":"Symbol", "TS":"Timestamp", "Day":"Day", "Open":"Open", "High":"High", "Low":"Low", "Close":"Close", "Volume":"Volume"}')
TBLPROPERTIES('mongo.uri'='mongodb://localhost:27017/marketdata.minibars');
... View more
03-31-2016
09:04 AM
@Ajay Kuppachi You need to use something like the below one . curl -u admin:admin -H "X-Requested-By: ambari" -X GET http://localhost:8080/api/v1/clusters/HDP_QA/services//FLUME_HANDLER curl -u admin:admin -H "X-Requested-By: ambari" -X DELETE http://localhost:8080/api/v1/clusters/HDP_QA/services/FALCON_SERVER here are the list of components names Ambari understands, [ { "name" : "KAFKA_BROKER" }, { "name" : "SUPERVISOR" }, { "name" : "METRICS_MONITOR" }, { "name" : "FLUME_HANDLER" }, { "name" : "SQOOP" }, { "name" : "YARN_CLIENT" }, { "name" : "FALCON_CLIENT" }, { "name" : "HDFS_CLIENT" }, { "name" : "OOZIE_CLIENT" }, { "name" : "YARN_CLIENT" }, { "name" : "ZOOKEEPER_CLIENT" }, { "name" : "HIVE_CLIENT" }, { "name" : "HCAT" }, { "name" : "MAPREDUCE2_CLIENT" }, { "name" : "HBASE_CLIENT" } ], { "name" : "NODEMANAGER" }, { "name" : "METRICS_MONITOR" }, { "name" : "DATANODE" }, { "name" : "SUPERVISOR" }, { "name" : "HBASE_REGIONSERVER" { "name" : "HBASE_MASTER" }, { "name" : "METRICS_MONITOR" }, { "name" : "ZOOKEEPER_SERVER" }, { "name" : "STORM_UI_SERVER" }, { "name" : "DRPC_SERVER" }, { "name" : "NIMBUS" }, { "name" : "WEBHCAT_SERVER" }, { "name" : "HIVE_SERVER" }, { "name" : "HIVE_METASTORE" }, { "name" : "TEZ_CLIENT" }, { "name" : "HCAT" } { "name" : "METRICS_MONITOR" }, { "name" : "METRICS_COLLECTOR" }, { "name" : "ZOOKEEPER_SERVER" }, { "name" : "HISTORYSERVER" }, { "name" : "APP_TIMELINE_SERVER" }, { "name" : "RESOURCEMANAGER" }, { "name" : "SECONDARY_NAMENODE" } { "name" : "NAMENODE" }, { "name" : "METRICS_MONITOR" }, { "name" : "ZOOKEEPER_SERVER" }
... View more
04-22-2016
07:23 PM
Actually, I checked that ExportSnapshot includes the Tools interface, hence it does support -files and -libjars so I think the approach is still good. Sorry, false panic.
... View more
03-29-2016
07:47 PM
4 Kudos
Hi @Vadim, OpenCV is famous for image processing in general. They have several tools for image and face recognition. Here is an example of how to do face recognition with OpenCV: tutorial. In terms of integration with Hadoop, there's a framework called HIPI developed by University of Virginia for leveraging HDFS and MapReduce for large scale image processing. This framework supports OpenCV too. Finally, for image processing in motion, you can use HDF with an OpenCV processor like the one published here
... View more
03-22-2017
04:27 AM
Hi Artem, I'm currently stuck in a particular use case where in I'm trying to access Hive Table data using spark.read.jdbc as shown below: export SPARK_MAJOR_VERSION=2 spark-shell import org.apache.spark.sql.{DataFrame, Row,SparkSession} val connectionProperties = new java.util.Properties() val hiveQuery = "(SELECT * from hive_table limit 10) tmp" val hiveResult = spark.read.jdbc("jdbc:hive2://hiveServerHostname:10000/hiveDBName;user=hive;password=hive", hiveQuery, connectionProperties).collect() But when I check for the results in hiveResult it's just empty. Could you please suggest what's going on here? I know we can access Hive tables using HiveSesssion and I've successfully tried that but is it possible to run hive queries and access Hive data using the above method?
... View more
12-27-2016
12:47 PM
Hi all,
Has anyone a workaround for this problem ? I have exactly the same case.
I have similar issues on the Sandbox 2.5 (VirtualBox-5.1.12-112440-Win - HDP_2.5_virtualbox).
I killed
the jobs with putty as root : yarn application -kill
application_1482410373661_0002 but they are still visible on Ambari.
[root@sandbox ~]# yarn application -kill application_1482410373661_0002
16/12/24 12:26:40 INFO impl.TimelineClientImpl: Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
16/12/24 12:26:40 INFO client.RMProxy: Connecting to ResourceManager at sandbox.hortonworks.com/172.17.0.2:8050
16/12/24 12:26:40 INFO client.AHSProxy: Connecting to Application History server at sandbox.hortonworks.com/172.17.0.2:10200
16/12/24 12:26:44 WARN retry.RetryInvocationHandler: Exception while invoking ApplicationClientProtocolPBClientImpl.getApplicationReport over null. Not retrying because try once and fail.
org.apache.hadoop.yarn.exceptions.ApplicationNotFoundException: Application with id 'application_1482410373661_0002' doesn't exist in RM.
I've found an issue corresponding :
Tez
client keeps trying to talk to RM even if RM does not know about the
application
https://issues.apache.org/jira/browse/TEZ-3156
This patch should be included as it was fixed for version 0.7.1
In the log (Ambary query) I can read 993 time :
INFO
: Map 1: 0/1 Reducer 2: 0/2
The query is the proposed in the tutorial : (
http://fr.hortonworks.com/hadoop-tutorial/hello-world-an-introduction-to-hadoop-hcatalog-hive-and-pig/#section_4)
<code>SELECT truckid, avg(mpg) avgmpg FROM truck_mileage GROUP BY truckid;
Any idea how to clear the history and restart without the running state ? Thanks in advance
... View more
03-26-2016
05:43 PM
1 Kudo
I found the problem. I must have missed the step "chown -R solr:solr /opt/lucidworks-hdpsearch/solr". Once I did this the query worked but still did not see tweets in the dashboard. I deleted the collection and reloaded it. After that data started to appear.
... View more
04-18-2016
01:32 PM
This error can also occur if the Atlas service is turned off. From the exception stack trace: at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136) Caused by: java.io.IOException: java.net.ConnectException: Connection refused at org.apache.atlas.security.SecureClientUtils$1$1.run(SecureClientUtils.java:107) at org.apache.atlas.security.SecureClientUtils$1$1.run(SecureClientUtils.java:99) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) at org.apache.atlas.security.SecureClientUtils$1.getHttpURLConnection(SecureClientUtils.java:99) at com.sun.jersey.client.urlconnection.URLConnectionClientHandler._invoke(URLConnectionClientHandler.java:159) at com.sun.jersey.client.urlconnection.URLConnectionClientHandler.handle(URLConnectionClientHandler.java:147) ... 26 more Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method) See the link here for more information: https://community.hortonworks.com/questions/22396/ranger-dependency-on-atlas.html
... View more
03-31-2016
09:46 PM
1 Kudo
You may want to try: https://hbasenet.codeplex.com/
... View more
03-29-2016
02:25 PM
atlast, am able to hit 127.0.0.1:8080 after several times deleting and importing VM's . in my case, the problem is am able to start ambari server , ambari agent ..but couldnot in web. i checked in windows command prompt whether it is listening to port 8080 by command 'netstat' ..it didnot listening so i added port 8080 in windows firewall advanced settings.. after starting ambari serevr..waited for few minutes and then hitted 127.0.0.1:8080 successsfully...
... View more