Support Questions

Find answers, ask questions, and share your expertise
Check out our newest addition to the community, the Cloudera Data Analytics (CDA) group hub.

ImportError: No module named numpy

Befor I post this issue, we have already readed all the same issue's solutions that we can find.


Our cluster is installed with cdh6.2, after install we use HUE to use the cluster. Job is submited via Hue.


When spark code need to import numpy,  got error below:


Traceback (most recent call last):
  File "/var/yarn/nm/usercache/admin/appcache/application_1557739482535_0001/container_1557739482535_0001_01_000001/", line 79, in <module>
    from import Vectors
  File "/var/yarn/nm/usercache/admin/appcache/application_1557739482535_0001/container_1557739482535_0001_01_000001/python/lib/", line 22, in <module>
  File "/var/yarn/nm/usercache/admin/appcache/application_1557739482535_0001/container_1557739482535_0001_01_000001/python/lib/", line 24, in <module>
  File "/var/yarn/nm/usercache/admin/appcache/application_1557739482535_0001/container_1557739482535_0001_01_000001/python/lib/", line 26, in <module>
ImportError: No module named numpy


We followed office guied to install anaconda parcel,  And setup the Spark Service Advanced Configuration Snippet (Safety Valve) for spark-conf/ 


export PYSPARK_PYTHON=/opt/cloudera/parcels/Anaconda/bin/python
export PYSPARK_DRIVER_PYTHON=/opt/cloudera/parcels/Anaconda/bin/python

Setup the Spark Client Advanced Configuration Snippet (Safety Valve) for spark-conf/spark-defaults.conf



Also, setup the YARN (MR2 Included) Service Environment Advanced Configuration Snippet (Safety Valve)



But non of these can help to solve the import issue.


Thanks for any help.




Please check if numpy is actually installed on all of the nodemanagers, if not, install it using below command (for python2.x) :


pip install numpy


If already installed, let us know the following: 


1) Can you execute the same command outside of hue i.e. using Spark2-submit ? Mention the full command here.

2) What spark command you use in Hue?

use command below, the job can be executed successfully.


export SPARK_HOME=/opt/cloudera/parcels/CDH/lib/spark
export HADOOP_CONF_DIR=/etc/alternatives/hadoop-conf
PYSPARK_PYTHON=/opt/cloudera/parcels/Anaconda/bin/python spark-submit --master yarn --deploy-mode cluster


In Hue,  open a spark snippet , select the py file, then run it. And the same code can also be executed in Hue's nodebook with yarn model.



We installed anaconda vir cdh. which is already actived.




In the below file:



we can see:


Screenshot from 2019-05-14 18-27-41.png



In a CDH 6.3.2 cluster have an Anaconda parcel distributed and activated, which of course has the numpy module installed. However the Spark nodes seem to ignore the CDH configuration and keep using the system wide Python from /usr/bin/python.


Nevertheless I have installed numpy in system wide Python across all cluster nodes. However I still experience the "ImportError: No module named numpy". Would appreciate any further advice how to solve the problem.


Not sure how to implement the solution referred in



@kernel8liang Could you please explain how to implement the solution?

Take a Tour of the Community
Don't have an account?
Your experience may be limited. Sign in to explore more.