我读了一些关于我现在看到的错误的文章,一些建议给,但是我已经看到了,但是我仍然看到了错误。1. create a yml file and include the needed packages (including the py4j)3. create a kernel pointing to the env5. running `import pyspark
_jdf.showString(n, int(truncate)))
/opt/SPARK2/lib/spark2/python/lib/py4j-0.10.4-src.zip/py4j/java_gateway.py/lib/spark2/python/lib/py4j-0.10.4-src.zip/py4j/protocol.py in get_return_value(ans
print sc.wholeTextFiles(data_dir).flatMap(update).top(10)#groupByKey().map(store) File "/home/user/spark2/python/pyspark/rdd.py", line 2107, in _jrddpickled_command = s
我在我的hortonworks集群上安装了Spark1.6.2和Spark2.0。每次启动spark-shell时,我都会得到:Multiple versions of Spark are installed but SPARK_MAJOR_VERSION is not set当我查看版本时,我得到:res0: String = 1.6.2
如何启动另一个版本(Spark2.0的火花外壳