Spark job fails in yarn-cluster mode -
my job runs in spark on yarn-client mode , fails on yarn-cluster mode, stating error "file not exist: hdfs://192.xxx.x.x:port/user/hduser/.sparkstaging/application_1442810383301_0016/pyspark.zip". although shows has uploaded file above directory!! cause??
this full error log:
application application_1449548654695_0003 failed 2 times due container appattempt_1449548654695_0003_000002 exited exitcode: -1000 more detailed output, check application tracking page:http://server1:8088/cluster/app/application_1449548654695_0003then, click on links logs of each attempt. diagnostics: file not exist: hdfs://192.168.0.109:54310/user/hduser/.sparkstaging/application_1449548654695_0003/pyspark.zip java.io.filenotfoundexception: file not exist: hdfs://192.168.0.109:54310/user/hduser/.sparkstaging/application_1449548654695_0003/pyspark.zip @ org.apache.hadoop.hdfs.distributedfilesystem$22.docall(distributedfilesystem.java:1309) @ org.apache.hadoop.hdfs.distributedfilesystem$22.docall(distributedfilesystem.java:1301) @ org.apache.hadoop.fs.filesystemlinkresolver.resolve(filesystemlinkresolver.java:81) @ org.apache.hadoop.hdfs.distributedfilesystem.getfilestatus(distributedfilesystem.java:1301) @ org.apache.hadoop.yarn.util.fsdownload.copy(fsdownload.java:253) @ org.apache.hadoop.yarn.util.fsdownload.access$000(fsdownload.java:63) @ org.apache.hadoop.yarn.util.fsdownload$2.run(fsdownload.java:361) @ org.apache.hadoop.yarn.util.fsdownload$2.run(fsdownload.java:359) @ java.security.accesscontroller.doprivileged(native method) @ javax.security.auth.subject.doas(subject.java:415) @ org.apache.hadoop.security.usergroupinformation.doas(usergroupinformation.java:1657) @ org.apache.hadoop.yarn.util.fsdownload.call(fsdownload.java:358) @ org.apache.hadoop.yarn.util.fsdownload.call(fsdownload.java:62) @ java.util.concurrent.futuretask.run(futuretask.java:262) @ java.util.concurrent.executors$runnableadapter.call(executors.java:471) @ java.util.concurrent.futuretask.run(futuretask.java:262) @ java.util.concurrent.threadpoolexecutor.runworker(threadpoolexecutor.java:1145) @ java.util.concurrent.threadpoolexecutor$worker.run(threadpoolexecutor.java:615) @ java.lang.thread.run(thread.java:745) failing attempt. failing application.
could provide full error log?
also did set master of application "yarn-cluster"? can python so:
conf = sparkconf().setappname(appname).setmaster("yarn-cluster") sc = sparkcontext(conf=conf)
Comments
Post a Comment