如何在pyspark中修复takeuptumleft stop迭代?

2024-05-18 08:43:49 发布

您现在位置:Python中文网/ 问答频道 /正文

Python3.7.0和spark 2.3.0有些奇怪。在

使用上述版本,我们得到以下例外:

19/08/10 21:29:10 WARN TaskSetManager: Lost task 0.0 in stage 6.0 (TID 13, 172.16.138.199, executor 1): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
  File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/rdd.py", line 1354, in takeUpToNumLeft
StopIteration

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/worker.py", line 229, in main
    process()
  File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/worker.py", line 224, in process
    serializer.dump_stream(func(split_index, iterator), outfile)
  File "/cvmfs/soft.computecanada.ca/easybuild/software/2017/Core/spark/2.3.0/python/lib/pyspark.zip/pyspark/serializers.py", line 372, in dump_stream
    vs = list(itertools.islice(iterator, batch))
RuntimeError: generator raised StopIteration

为什么会有上述例外?怎么解决呢?在


Tags: inpycoreliblinesoftwarezipspark