Member since
03-16-2016
707
Posts
1753
Kudos Received
203
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 6967 | 09-21-2018 09:54 PM | |
| 8724 | 03-31-2018 03:59 AM | |
| 2615 | 03-31-2018 03:55 AM | |
| 2754 | 03-31-2018 03:31 AM | |
| 6176 | 03-27-2018 03:46 PM |
12-28-2016
04:28 PM
@Ashnee Sharma There was an issue and for that you submitted a question separately. It is good to document here as well, for other sake that may be encounter a similar problem. Please post it. I found it. Based on the original response, you encountered an issue, then you asked this question: https://community.hortonworks.com/questions/74245/how-to-disable-pagination-for-ambari-ldap.html
... View more
12-28-2016
03:53 PM
2 Kudos
@Brad Bukacek Jr By design, the HBase REST server returns content encoded response with base64. So all your content, like the column family, the qualifier and the raw content will be encoded. You just need to create a custom JSON deserializer. Here is an awesome blog about this subject: https://blog.layer4.fr/2016/11/16/hbase-rest-api-knox-java/ There is a special section about your problem.
... View more
12-27-2016
12:50 AM
See link below to learn why s3a is a better option than s3n, but that may not be the cause for your issue. https://wiki.apache.org/hadoop/AmazonS3
... View more
12-26-2016
10:50 PM
1 Kudo
@Dmitry Otblesk Login to Ambari UI first then click on YARN link on the left nav bar then on the QuickLinks and chose Resource Manager UI link. You could also go directly to your Resource Manager UI if you know the host where the Resource Manager service runs also the port. You should also take advantage of Hive Tez View to see all the tasks executed and time needed for each. While you execute the query watch the execution in Resource Manager UI to understand number of containers per task, resource utilization etc. If you see that you have low degree of parallelism and still resources enough to instantiate more containers then you have an opportunity to adjust the query to allow more parallelism. http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_performance_tuning/content/ch_query_optimization_hive.html Change the path from the link to match your version of HDP.
... View more
12-26-2016
10:38 PM
2 Kudos
@Simran Kaur You need to use a join or a sub-query within the sub-query. You can access out query from inner query. It is the other way around and by design for any SQL-like language.
... View more
12-26-2016
10:34 PM
2 Kudos
@Fish Berh This could have due to a problem with the spark-csv jar. i have encountered this myself and I found a solution which I cannot find now. Here are my notes at the time: 1. Create a folder in your local OS or HDFS and place the proper versions for your case of the jars here (replace ? with your version needed):
spark-csv_?.jar commons-csv-?.jar univocity-parsers-?.jar 2. Go to your /conf directory where you have installed Spark and in the spark-defaults.conf file add the line: spark.driver.extraClassPath D:/Spark/spark_jars/* The asterisk should include all the jars. Now run Python, create SparkContext, SQLContext as you normally would. Now you should be able to use spark-csv as sqlContext.read.format('com.databricks.spark.csv').\
options(header='true', inferschema='true').\
load('foobar.csv')
... View more
12-26-2016
10:25 PM
2 Kudos
@kishore sanchina Your spark use must be able to create folder under that /tmp/spark-tmp. Based on your comments you did not grant ownership successfully. You should grant recursive ownership of /tmp as such that it will include all subfolders existent or created at runtime: chown spark -R /tmp I assumed your user is spark. However, I really don't like the idea of using /tmp for that (SA taste).You should use maybe a folder created under SPARK_HOME.
... View more
12-26-2016
10:09 PM
2 Kudos
@Timothy Spann Added @Chris Nauroth to the thread. He is a mentor in this Apache project.
... View more
12-26-2016
10:01 PM
2 Kudos
@Raghvendra Singh Tutorial: http://hortonworks.com/hadoop-tutorial/getting-started-with-pivotal-hawq-on-hortonworks-sandbox/ Look for section USING OTHER TOOLS TO WORK WITH HAWQ Follow instructions on how to download ODBC/JDBC driver and how to use it. If your data stored is JSON then you are set, otherwise you have to handle it before displaying to your d3js based dashboard. HAWQ is SQL-like database with advance ANSI compliance.
... View more
12-26-2016
09:55 PM
1 Kudo
@Manoj Ramakrishnan You should try to reduce the heap to the minimum required and also use g1gc. This is due to a GC tuning issue.
... View more