Python3.7.0和spark 2.3.0有些奇怪。在
使用上述版本,我们得到以下例外:
19/08/10 21:29:10 WARN TaskSetManager: Lost task 0.0 in stage 6.0 (TID 13, 172.16.138.199, executor 1): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/rdd.py", line 1354, in takeUpToNumLeft
StopIteration
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/worker.py", line 229, in main
process()
File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/worker.py", line 224, in process
serializer.dump_stream(func(split_index, iterator), outfile)
File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/serializers.py", line 372, in dump_stream
vs = list(itertools.islice(iterator, batch))
RuntimeError: generator raised StopIteration
为什么会有上述例外?怎么解决呢?在
目前没有回答
相关问题 更多 >
编程相关推荐