Member since
06-07-2016
8
Posts
2
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
37361 | 06-10-2016 08:23 AM |
03-27-2017
08:23 AM
I would like to use Pythons Logging library, but want the output of the logs to land in HDFS instead of the local file system for the worker node. Is there a way to do that? My code for setting up logging is below: import logging logging.basicConfig(filename='/var/log/DataFramedriversRddConvert.log',level=logging.DEBUG) logging.basicConfig(format='%(asctime)s %(message)s') logging.info('++++Started DataFramedriversRddConvert++++')
... View more
Labels:
- Labels:
-
Apache Spark
-
HDFS
03-02-2017
04:52 AM
Can you clarify your question? Do you mean regarding jupyter or what the core was from?
... View more
10-17-2016
02:18 PM
Ah, my error was not using HDFS: for the .py. Thanks!
... View more
10-17-2016
01:12 PM
I'm getting this error also. Have you managed to solve it?
... View more
06-10-2016
08:23 AM
1 Kudo
Solved. I had to: $ export PATH= /opt/cloudera/parcels/Anaconda-4.0.0/bin:$PATH I had Python installed before installing Anaconda, and it was clashing.
... View more
06-08-2016
09:46 AM
1 Kudo
I'm trying to install Jupyter using these instructions http://www.cloudera.com/documentation/enterprise/5-6-x/topics/spark_ipython.html. I've have installed under Anaconda and when I try to run pyspark ( from section
Starting a Notebook with PySpark) I receive the following error:
$ pyspark
Traceback (most recent call last):
File "/opt/cloudera/parcels/Anaconda/bin/jupyter", line 4, in <module>
from jupyter_core.command import main
ImportError: No module named jupyter_core.command
Any help is appreciated.
... View more
Labels:
- Labels:
-
Cloudera Enterprise Data Hub