Zohaa Qamar
01/29/2023, 3:57 AM# java.lang.OutOfMemoryError: Java heap space
# -XX:OnOutOfMemoryError="kill -9 %p"
# Executing /bin/sh -c "kill -9 27571"...
I have tried using various instance types like 20-25 core instances of m5.16xlarge and r5.12xlarge. Also, tried playing around spark configurations like spark.driver.memory and spark.executor.memory from 30g to 300g, but nothing helped. The job has not any major computation but simply its spark.coalesce.write.partitionBy.parquet. Also tried setting _HADOOP_HEAPSIZE_ in the configuration to 100g and 200g. Please ask if more information is required. Thanks. Here is the screenshot of EMR executors:JosephK (exDatabricks)
01/29/2023, 10:47 AMZohaa Qamar
01/29/2023, 11:04 PMspark.driver.memory=100g
spark.executor.memory=100g
spark.yarn.executor.memoryOverhead=2g
spark.yarn.driver.memoryOverhead=2g
spark.executor.cores=5
spark.driver.cores=5
spark.executor.instances=60
spark.network.timeout=50000s
spark.executor.heartbeatInterval=5000s
But still it is failing in EMR. I have the below errors now:
ERROR AsyncEventQueue: Dropping event from queue eventLog. This likely means one of the listeners is too slow and cannot keep up with the rate at which tasks are being started by the scheduler.
and also this in stderr:
java.lang.NullPointerException: null value for spark.executor.cores
I have tried with several variations in the spark configurations and coalesce/repartition etc. But there is no way it is succeeding. Any help please?Mike M
01/30/2023, 3:49 AMZohaa Qamar
01/30/2023, 3:59 AMsparkSession.read().schema(getInputSchema()).json(sourceFilesPath)
Mike M
01/30/2023, 4:00 AMZohaa Qamar
01/30/2023, 4:01 AMMike M
01/30/2023, 4:03 AMZohaa Qamar
01/30/2023, 4:03 AMMike M
01/30/2023, 4:04 AMZohaa Qamar
01/30/2023, 4:06 AMMike M
01/30/2023, 4:12 AMZohaa Qamar
01/30/2023, 4:13 AMMike M
01/30/2023, 4:17 AMspark.sql.shuffle.partitions
value?Zohaa Qamar
01/30/2023, 4:24 AMMike M
01/30/2023, 4:30 AMZohaa Qamar
01/30/2023, 4:33 AMMike M
01/30/2023, 4:37 AMZohaa Qamar
01/30/2023, 4:40 AMJosephK (exDatabricks)
01/30/2023, 12:25 PMZohaa Qamar
01/30/2023, 3:38 PMbreaking it up in to smaller jobs may help
does this mean I process my job for chunks of data multiple times?JosephK (exDatabricks)
01/30/2023, 3:56 PMZohaa Qamar
01/30/2023, 5:35 PMJosephK (exDatabricks)
01/30/2023, 5:42 PMZohaa Qamar
01/30/2023, 5:43 PMJosephK (exDatabricks)
01/30/2023, 5:45 PMZohaa Qamar
01/30/2023, 5:47 PM