I need to run Apace Spark script on Amazon EC2. Script uses such libs, as numpy, pandas etc. The trouble is that I have numpy installed in
/usr/local/lib64/python2.7/site-packages, and this folder isn't in PYTHONPATH by default. So when I export PYTHONPATH=$PYTHONPATH:/usr/local/lib64/python2.7/site-packages, usual python detects it (import numpy causes no problems), but when I'm trying to import it in pyspark shell - it shows:
>>> import numpy
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: No module named numpy
>>> exit()
is there any solution how to change pyspark's PYTHONPATH?