Member since
05-09-2016
280
Posts
58
Kudos Received
31
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3762 | 03-28-2018 02:12 PM | |
3024 | 01-09-2018 09:05 PM | |
1650 | 12-13-2016 05:07 AM | |
5062 | 12-12-2016 02:57 AM | |
4325 | 12-08-2016 07:08 PM |
11-28-2016
11:36 PM
Using Hortonworks Sandbox, I am setting up SparkR in both RStudio and Zeppelin. This below code works properly in RStudio and SparkR shell but not in Zeppelin, please have a look: if (nchar(Sys.getenv("SPARK_HOME")) < 1) {
Sys.setenv(SPARK_HOME = "/usr/hdp/2.5.0.0-1245/spark")
}
library(SparkR, lib.loc = c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib")))
sc <- sparkR.init(master = "local[*]", sparkEnvir = list(spark.driver.memory="2g"),sparkPackages="com.databricks:spark-csv_2.10:1.4.0")
sqlContext <- sparkRSQL.init(sc)
train_df <- read.df(sqlContext,"/tmp/first_8.csv","csv", header = "true", inferSchema = "true") But when I do this in Zeppelin using livy.spark interpreter, I get ClassNotFound Exception: java.lang.ClassNotFoundException: Failed to find data source: csv. Please find packages at http://spark-packages.org I am also importing the dependencies using dep interpreter - %dep
z.reset()
z.load("com.databricks:spark-csv_2.10:1.4.0") But this seems to make no impact I guess. I have also tried manually copying spark-csv_2.10-1.4.0.jar to /usr/hdp/2.5.0.0-1245/spark/lib, but it is not working. Has anyone experienced this before? Thanks in advance
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache Zeppelin
11-27-2016
10:45 PM
is it the right solrconfig.xml file?
... View more
11-27-2016
10:35 PM
1 Kudo
@Raf Mohammed , make sure you edit the right solrconfig.xml file. As per the tutorial, you have to edit the solrconfig.xml of tweet_configs folder (/opt/lucidworks-hdpsearch/solr/server/solr/configsets/tweet_configs/conf/solrconfig.xml). After you open this file via vi shell, add <str>EEE MMM d HH:mm:ss Z yyyyy</str> inside section solr.ParseDateFieldUpdateProcessorFactory
... View more
11-27-2016
09:08 PM
@rishabh jain, go to Resource Manager UI and kill the running applications. Your query might not be getting the required resources, then run the query again.
... View more
11-27-2016
07:14 PM
@Raf Mohammed, with that permission,you should be able to edit the file. Can you please open that file? vi /opt/lucidworks-hdpsearch/solr/server/solr/configsets/data_driven_schema_configs_hdfs/conf/solrconfig.xml
... View more
11-27-2016
06:54 PM
@Raf Mohammed
You ran chown command from root user, if you would have been ran it from solr user, that would have worked too. Running from root basically wasted your usermod thing. Coming to the current issue, can you please check the permission of that file you want to access.
... View more
11-27-2016
05:59 PM
1 Kudo
@Raf Mohammed , you can change ownership file if you're root. You cannot use sudo fo solr user because it does not have an sudo access. If you run sudo chown -R solr:solr /opt/lucidworks-hdpsearch/solr, it will not work. You can however change the owning group to a group you're a member of. Your directory has solr:hadoop (as user and group) and you are changing it to solr:solr. But if you check whether solr group has solr user, you will not be able to find it because it is not there. So now you have to add solr user to the solr group. Run this command as root: usermod -a -G solr solr Now if you try your command, it should work. PS: Doing this : sudo -u hdfs hadoop fs chown -R solr:solr /opt/lucidworks-hdpsearch/solr does not make any sense because it is a local directory, not HDFS directory.
... View more
11-27-2016
04:12 PM
Hi @Brice LAM VAN , zeppelin view has been removed from the Sandbox, we will make that update in the tutorial. For the meantime, you can access Zeppelin UI on 127.0.0.1:9995
... View more
11-27-2016
01:08 AM
Navigate to Resource Manager UI on 127.0.0.1:8088. Click on Running applications and then kill them. You will see a button to Kill the application. After this, run the same query again.
... View more