Spark Python QA
Spark Python QA
Q: PySpark: java.lang.OutofMemoryError: Java heap space
1 | PySpark: java.lang.OutofMemoryError: Java heap space |
A
1 | spark_conf.setAppName("recommend").setMaster("local[*]").set('spark.executor.memory', '12g').set('spark.driver.memory', '14g') |
Q: Please install psutil to have better support with spilling
1 | UserWarning: Please install psutil to have better support with spilling |
A
1 | pip install psutil |
Q: {0}.{1} does not exist in the JVM
1 | "{0}.{1} does not exist in the JVM".format(self._fqn, name)) |
A
1 | pyspark 与spark版本不对 |
Q: Python worker failed to connect back
1 | 22/06/16 12:20:00 WARN TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1) (192.168.3.96 executor 0): org.apache.spark.SparkException: Python worker failed to connect back. |
A
1 | PYSPARK_PYTHON=D:\ProgramData\Anaconda3\python.exe |
All articles on this blog are licensed under CC BY-NC-SA 4.0 unless otherwise stated.