Member since
03-25-2016
142
Posts
48
Kudos Received
7
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5774 | 06-13-2017 05:15 AM | |
1903 | 05-16-2017 05:20 AM | |
1341 | 03-06-2017 11:20 AM | |
7853 | 02-23-2017 06:59 AM | |
2223 | 02-20-2017 02:19 PM |
02-23-2017
11:03 AM
@Oriane Try option "Reward User"
... View more
02-23-2017
10:54 AM
Thanks @Jay SenSharma
... View more
02-23-2017
10:51 AM
@Oriane I am glad you have this working now. If you believe I helped, please vote up my answer and select as best one 🙂
... View more
02-23-2017
10:26 AM
@Oriane For the "prefix not found" - double-check if you have spark interpreter binded in that notebook. See my screenshot - Spark needs to be "blue"
... View more
02-23-2017
09:39 AM
@Oriane Do exactly this: - in the new section type: %spark - press <Enter> button - type: sc.version - press <Enter> button Now, run it Does this help? I am asking as noticed that the copied code causing issues.
... View more
02-23-2017
06:59 AM
2 Kudos
@Oriane Can you provide the following: 1. As @Bernhard Walter already asked, can you attach the screenshot of your spark interpreter config from Zeppelin UI 2. Create a new Notebook and run the below and send the output: %sh
whoami 3. Can you attach the output of $ ls -lrt /usr/hdp/current/zeppelin-server/local-repo 4. Is your cluster Kerberized?
... View more
02-20-2017
02:19 PM
1 Kudo
OK - I get this working now. If anyone interested, here you are: val df = sqlContext.sql("SELECT * from table1")
val tempResult = df.filter(df("field1") > 10)
tempResult.write.mode("overwrite").saveAsTable("default.new_table")
val df1 = sqlContext.sql("SELECT * from default.new_table")
df1.show()
NOTE: the "new_table" table can but does not need to exist before writing to it
... View more
02-20-2017
11:28 AM
4 Kudos
Problem Interpreters do not work through zeppelin with no internet access. Checking out /usr/hdp/current/zeppelin-server/local-repo - the folder is empty or contains “org” folder with no jars Running hive through Zeppelin returns %jdbc(hive)
show tables;
...
org.apache.hive.jdbc.HiveDriver class java.lang.ClassNotFoundException
... Solution The permanent solution is currently planned to be delivered in HDP 2.6. Here is a workaround to follow: a) I have tarred /usr/hdp/current/zeppelin-server/local-repo into zeppelin-local-repo.tar.gz and located on https://drive.google.com/drive/folders/0B-YVWxQz56HubWhUdEdVWGZ1Mms?usp=sharing. As this is my google drive I can allow you to access it after receiving the request from you b) download the file into /tmp/zeppelin folder c) extract it - this should created local-repo folder. So, you will get /tmp/zeppelin/local-repo with all the subfolders in it d) copy the content of /tmp/zeppelin/local-repo into /usr/hdp/current/zeppelin-server/local-repo e) change the owner of local-repo and all its folders/file to zeppelin:hadoop (or zeppelin:hdfs <- whatever the GROUP you have) f) change the permissions local-repo and all its folders/file to 755 g) restart Zeppelin service NOTE: step a) is something you can do yourself. Just install the environment on a temporary machine with internet access and get the content of /usr/hdp/current/zeppelin-server/local-repo from there.
... View more
Labels:
02-20-2017
08:52 AM
1 Kudo
I have got the following: val df = sqlContext.sql("SELECT * from table1")
var tempResult = df.filter(df("field1") > 10) I have also already created another table - table2 - with the same structure as table1. How can I save/insert the result of tempResult into table2?
... View more
Labels:
- Labels:
-
Apache Spark
02-14-2017
03:49 PM
1 Kudo
@Srikanth Puli Please, have a look at this: https://cwiki.apache.org/confluence/display/Hive/StatsDev#StatsDev-ExistingTables–ANALYZE For a non-partitioned table, you can issue the command: ANALYZE TABLE Table1 COMPUTE STATISTICS FOR COLUMNS; to gather column statistics of the table (Hive 0.10.0 and later). If Table1 is a partitioned table, then for basic statistics you have to specify partition specifications like above in the analyze statement. Otherwise a semantic analyzer exception will be thrown. However for column statistics, if no partition specification is given in the analyze statement, statistics for all partitions are computed. You can view the stored statistics by issuing the DESCRIBE command. Statistics are stored in the Parameters array. Suppose you issue the analyze command for the whole table Table1, then issue the command: DESCRIBE EXTENDED TABLE1; then among the output, the following would be displayed: ... , parameters:{numPartitions=4, numFiles=16, numRows=2000, totalSize=16384, ...}, .... I hope this helps.
... View more