Member since
09-29-2015
155
Posts
205
Kudos Received
18
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
8524 | 02-17-2017 12:38 PM | |
1368 | 11-15-2016 03:56 PM | |
1914 | 11-11-2016 05:27 PM | |
15617 | 11-11-2016 12:16 AM | |
3139 | 11-10-2016 06:15 PM |
08-24-2016
01:39 PM
@Alexander is there a full list of these hdi scripts available? If not how did you discover the ones above?
... View more
08-24-2016
01:25 PM
3 Kudos
+1 on Recommender system. A more concrete example is "Building a Movie Recommendation Service with Apache Spark" below that walks you through an example. https://www.codementor.io/spark/tutorial/building-a-recommender-with-apache-spark-python-example-app-part1
... View more
08-23-2016
04:16 PM
It seems the new version of Sandbox does not have R pre-installed. Its an easy installation procedure: sudo yum install -y epel-release
sudo yum update -y
sudo yum install -y R
... View more
08-23-2016
04:15 PM
It seems the new version of Sandbox does not have R pre-installed. Its an easy installation procedure : sudo yum install -y epel-release
sudo yum update -y
sudo yum install -y R
... View more
08-02-2016
04:18 PM
You cant have another zeppelin notebook service deployed using Ambari. You will be able to have multiple services like Zeppelin in upcoming release, but its not a GA function right now.
... View more
08-01-2016
07:51 PM
@wsalazar This is what i use to build the zeppelin: mvn clean package -Pspark-1.X -Dhadoop.version=2.X.0 -Phadoop-2.X -Pyarn -DskipTests swap .X for the version you are building against. It makes it YARN compatible.
... View more
08-01-2016
07:42 PM
2 Kudos
@henryon wen You definitely can, i built the apache zeppelin from source code like this: checkout the zeppelin project: git clone https://github.com/apache/incubator-zeppelin and than build, make sure u have mvn installed. mvn clean package -Pspark-1.X -Dhadoop.version=2.X.0
-Phadoop-2.X -Pyarn -DskipTests
... View more
07-20-2016
10:35 PM
4 Kudos
Greetings Experts, I am trying to run LivyServer in zeppelin in HDP 2.5TP sandbox, however my interpreter keeps failing. I tried %livy.sparkr and %livy.pyspark all getting same exceptions. See error messages below from the livy log file: LivyPySparkInterpreter while interpret
java.lang.Exception: Cannot start pyspark
INFO [2016-07-18 22:20:23,102] ({Thread-0} RemoteInterpreterServer.java[run]:81) - Starting remote interpreter server on port 41012
INFO [2016-07-18 22:20:23,420] ({pool-1-thread-4} RemoteInterpreterServer.java[createInterpreter]:169) - Instantiate interpreter org.apache.zeppelin.livy.LivySparkR
Interpreter
INFO [2016-07-18 22:20:23,457] ({pool-1-thread-4} RemoteInterpreterServer.java[createInterpreter]:169) - Instantiate interpreter org.apache.zeppelin.livy.LivySparkI
nterpreter
INFO [2016-07-18 22:20:23,461] ({pool-1-thread-4} RemoteInterpreterServer.java[createInterpreter]:169) - Instantiate interpreter org.apache.zeppelin.livy.LivyPySpar
kInterpreter
INFO [2016-07-18 22:20:23,462] ({pool-1-thread-4} RemoteInterpreterServer.java[createInterpreter]:169) - Instantiate interpreter org.apache.zeppelin.livy.LivySparkS
QLInterpreter
INFO [2016-07-18 22:20:23,505] ({pool-2-thread-2} SchedulerFactory.java[jobStarted]:131) - Job remoteInterpretJob_1468880423503 started by scheduler org.apache.zepp
elin.livy.LivySparkRInterpreter1573372211
ERROR [2016-07-18 22:20:24,722] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is starting
ERROR [2016-07-18 22:20:25,754] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is starting
ERROR [2016-07-18 22:20:26,769] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is starting
ERROR [2016-07-18 22:20:27,784] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is dead
ERROR [2016-07-18 22:20:28,799] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is dead
ERROR [2016-07-18 22:20:29,814] ({pool-2-thread-2} LivyHelper.java[createSession]:89) - sessionId:0.0 state is dead
....
....
ERROR [2016-07-18 22:41:10,602] ({pool-2-thread-6} LivyHelper.java[createSession]:89) - sessionId:2.0 state is starting
ERROR [2016-07-18 22:41:11,609] ({pool-2-thread-6} LivyHelper.java[createSession]:114) - Error getting session for user within 60Sec.
ERROR [2016-07-18 22:41:11,609] ({pool-2-thread-6} LivyHelper.java[createSession]:120) - Error getting session for user
java.lang.Exception: Cannot start pyspark.
at org.apache.zeppelin.livy.LivyHelper.createSession(LivyHelper.java:115)
at org.apache.zeppelin.livy.LivyPySparkInterpreter.interpret(LivyPySparkInterpreter.java:63)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:94)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:341)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR [2016-07-18 22:41:11,610] ({pool-2-thread-6} LivyPySparkInterpreter.java[interpret]:70) - Exception in LivyPySparkInterpreter while interpret
java.lang.Exception: Cannot start pyspark.
at org.apache.zeppelin.livy.LivyHelper.createSession(LivyHelper.java:115)
at org.apache.zeppelin.livy.LivyPySparkInterpreter.interpret(LivyPySparkInterpreter.java:63)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:94)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:341)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
INFO [2016-07-18 22:41:11,611] ({pool-2-thread-6} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1468881610078 finished by scheduler org.apache.ze
ppelin.livy.LivyPySparkInterpreter1321637285
... View more
Labels:
- Labels:
-
Apache Spark
06-14-2016
05:56 PM
For full guide on the ambari quickstart on vagrant follow the apache doc: https://cwiki.apache.org/confluence/display/AMBARI/Quick+Start+Guide
... View more
05-04-2016
03:22 AM
@Vadim will give it a try tomorrow, sorry for delay in response
... View more