Manual submitted step jobs failing

0

We have a airflow setup runs the EMR jobs daily basis. I noticed an odd behavior that when I resubmit job for calculating the adhoc reports, spark application failed with below error, arguments seems to be leading some issue at AWS end as the same job worked before.

" No main class set in JAR; Please specify one with --class "

Vaas
질문됨 6달 전197회 조회
1개 답변
3
수락된 답변

Hello,

Sorry for the inconvenience caused. This is an ongoing issue in EMR Step console for spark-application type as the order of the arguments changed when submitting. Internal team is being worked on fixing this issue and you will be able to submit the jobs as it was before once fixed. Meanwhile, to unblock from this situation, you can do workaround with below,

  1. Submit the job through command-runner and pass the spark-submit as argument,

    a. Choose "Customized Jar" in Add step

    b. Add "command-runner.jar" in the Jar-Location

    c. In the Argument field, place the entire spark-submit command (example - spark-submit --deploy-mode cluster --class org.apache.spark.examples.SparkPi s3:///samplejob/spark-examples.jar 10)

  2. Use CLI method,

aws emr add-steps --cluster-id <clusterid> --steps Type=Spark,Name="SparkPi",Args=[--deploy-mode,cluster,--class,org.apache.spark.examples.SparkPi,s3://<s3bucket>/samplejob/spark-examples.jar,10],ActionOnFailure=CONTINUE

AWS
지원 엔지니어
답변함 6달 전
profile picture
전문가
검토됨 6달 전
profile pictureAWS
전문가
검토됨 6달 전
  • Thank you very much for the prompt response. I am able to submit the job with suggested alternatives. Hope this issue will be fixed asap.

로그인하지 않았습니다. 로그인해야 답변을 게시할 수 있습니다.

좋은 답변은 질문에 명확하게 답하고 건설적인 피드백을 제공하며 질문자의 전문적인 성장을 장려합니다.

질문 답변하기에 대한 가이드라인

관련 콘텐츠