AWS EMR [System Error] Fail to delete the temp folder

0

In AWS EMR, I encountered the following error message when running a pyspark job, which ran successfully on my local machine.

[System Error] Fail to delete the temp folder

Is there a way to troubleshoot this? Is this a permissions issue with the temp folder access by EMR, accessible across all jobs?

已提问 5 个月前235 查看次数
1 回答
4
已接受的回答

Hello,

Yes, it looks like either the permission issue or the tmp files might be in use.

  1. Please check if you try opening the pyspark shell as hadoop user in EMR or use sudo pyspark in hadoop user
  2. Try checking the spark-shell working without issue instead of pyspark shell.
  3. Include spark.local.dir to different local directory in primary node to see if this fix the issue.
  4. Restart the spark service(sudo systemctl restart spark-history-server.service
  5. Set the log level to debug rootLogger.level = debug in log4j file /etc/spark/conf/log4j2.properties and retry the pyspark shell. This might give more insight to understand the issue.
AWS
支持工程师
已回答 5 个月前

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则

相关内容